var/home/core/zuul-output/0000755000175000017500000000000015150542073014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015150553531015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000204043015150553353020256 0ustar corecore֢ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs$r.k9Gf8^PF]u|l&U狿li//|y-o\׼ٮ֜V˿l oo-q3f2f(_}^ſq^?-}ԏ|W E^ <]]ϗj;Hw70g"Gǯ/7ݮfGR)$DD D~m\rٲ] g $\xeK49)ol7(ӗofՈ_k0mZpPneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZwR}@E1.aLk>0M.*x6Q#%q^Hjf4}&|dd#)3c 0'Iw A57&Q"ԉQIF$%* tJ,w˼љ:72 2Zi}e0ʡPBU[.z{&:ɴ@=yN,:;zuL{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}[^Sgh+NqɌDSd1d9nT#Abn q1J# !8,$RNI?]\EZ~^- o^nNz&BBbm1lχtO Ң`?Tӣ 5W=Xzߙoux5n|X&pi]>w5p !Ɛ%aޗ' B.-^ mQYd'xP2cyڈL|Z΢rZg7n͐AG%ʷr<8 2S>h?y| (GClsXT(VIx$(J:&~,e 1ķ/5^Z-y`)͐-o΁qGWo(C U ?}aK+,&?>Y;ufʕ"uZ0EyT0: =XTy-nhI׼&q]#v0nFNV-9JϲdK\D2s&[#bE(mV9ىb_ȾcmXQj#3hEEH*Of äE@O0~yot3iYhZW~^ 9^oķ[Z`~}PSSii4ȷT (Dn@6_V3E炱}r4(9izh38u'8KwI~3v4&8[qߏ5.)Q VE JN`:!W5oӿ":7erԮoQ#% H!PK)~EZ{P+-/< )!M)aUrC|}_~ܥ0~fgkԇAt8>uJڡcp4A{C U =TCȾ39h®3uO0T-Oe+tǭQI%Q$SiJ. 9F[L1c!zG|k{kEu+Q & "X@+.+s~t?53!}~V[F)RH?uvͪ Ǐ6'6)уGہwQ5$qI[|AbH&Y``:"s ayiBq)u%'4 yܽ y_0 -i̭uJ{KưЖ@+UBj -&JO y@}D=S.€>3T0|9ē7$3Érhx^-$Zøu32EʉD'MS3}t i:Y`cФIX0$3VQ 7,oTWNJOAqO:rƭĘ DuZ^ To3dEN/}fq%G*YSUZa{P,97óI,Q{eNFV+(hʺb 7ʞXlýcsT 'zXm|oI7cGԓL!$<[yҘJqܚ%G)|A{+V#dFjVh#um`;Duhc=JS\kkZAY`] s Exb 5%woR[oI]${&L8<=#0yaKL: JJl rw#H+B|ɧJiM c/;m#NS80J=l}._?M<lu Y> XH\z:dHElL(uHR0i#q%]K&>.~y7/(.Xݚi;n%PAU?/(z_n}9U}'/osVu[H<9˷dDLF amKGm+`ùXz!"ܚ: hL\1u޷P_U{GwNNXYt\oq_@gOV ]cӰJ:^q';{wV=mdZB4']a.QO:#'6RE'E3 */HAYk|z|ؾPQgOiMcĚ$H4x>bl=pd9YfAMpIrv̡}XIթJ:|nl^/GSZ;m#Nvj{,4xPA*1bv^JLj&DY3#-1*I+g8a@(*%kޏ=S-ݑ2ƹڞ7կZa{0dqȾ3̗w.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t?}d,ZByX/&Ksg3["66hŢFD&iQCFd4%h=RA?ed}ĥ6ui ; dFF끄߽c1e-7C5_7XѪ;Ʃ2tStΆ,~Lp`-;uIBqBVlU_~F_v+ERz#{)@o\!@q['&&$"THl#d0 %L+`8zOҚƞ`wF~dWCg5o|ΔC1s`u.EkB6ga׬9J2{?+Fi7Z(ZN~;MM/yu2CݬP]Wtɖ9F.[-2, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTn!tT̅Rhɇ ќuޏ¢6}#LpFD58LQ Lf~/EOF p rH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{B.|!p+,ICE^fu `|M3J#BQȌ6DNnCˣ"F$/Qx%m&FK_7P|٢?I-RiAKoQrMI>QQ!'7h,sF\jzP\7:Q\)#s{p'ɂN$r;fVkv߸>6!<̅:xn<# -BȢ1I~ŋ-*|`В~_>ۅm}67X9z=Oa Am]fnޤ{"hd߃Ԉ|tLD3 7'yOc& LFs%B!sRE2K0p\0͙npV)̍F$X8a-bp)5,] Bo|ؖA]Y`-jyL'8>JJ{>źuMp(jL!M7uTźmr(Uxbbqe5rZ HҘ3ڴ(|e@ew>w3C=9k-{p>րd^T@eFZ#WWwYzK uK r؛6V L)auS6=`#(TO֙`mn Lv%7mSU@n_Vۀl9BIcSxlT![`[klzFض˪.l >7l@ΖLl gEj gWUDnr7AG;lU6ieabp៚U|,}S@t1:X _ .xI_7ve Z@7IX/C7@u BGڔE7M/k $q^hڧ};naU%~X!^C5Aw͢.@d!@dU}b? -ʏw |VvlK۴ymkiK% 0OFjT_kPW1mk%?\@R>XCl}b ,8; :.b9m]XaINE`!6uOhUuta^xN@˭d- T5 $4ذ:[a>֋&"_ }Oõϸ~rj uw\h~M il[ 2pCaOok.X0C?~[:^Pr򣏷y@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ  l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7#`VCpᇽmpM+tWuk0 q /} 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7M'#|kf1:X l]ABC {kanW{ 6 g`_w\|8Fjȡstuf%Plx3E#zmxfU S^ 3_`wRY}@ŹBz²?mК/mm}m"Gy4dl\)cb<>O0BďJrDd\TDFMEr~q#i}$y3.*j) qQa% |`bEۈ8S 95JͩA3SX~߃ʟ~㍖›f!OI1R~-6͘!?/Vvot4~6I@GNݖ-m[d<-l9fbn,'eO2sٟ+AWzw A<4 }w"*mj8{ P&Y#ErwHhL2cPr Wҭюky7aXt?2 'so fnHXx1o@0TmBLi0lhѦ* _9[3L`I,|J @xS}NEij]Qexx*lJF#+L@-ՑQz֬]")JC])"K{v@`<ۃ7|qk" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ"0(HKkD4<80: M:'֥P!r "Lӓݰ@ 9n# " $fGgKQӦ4}Gn\^=-Y5PI dPN6 Ozځ/פ|5) F[ڣ$2*%&h v%9HN H~Q+oi?&۳)-nqK?2ސv/3,9ҮT9Cef˝49i.2DxatC<8iR/ƬйR֌vN8J"iJ. T>)qaY4ͬlyg "]BvW#99`TegõII kюHLa^c&/H^FFIu`2a$mc Ry+R:LڕDܓ>Y:]t.+|PT6=qWe0NƏw<6o3mv8k vGOfpEOkÈWȤMف lOc;SR&.w,qk>MPs+Xh4iyuGRd֞q鮺]m S{}]U kV0/ŜxtADx"Xh4|;XSxߵă@pE:y]/"(MCG`ʶϊGi+39#gNZYE:Qw9muB`9`LDhs4Ǩ9S`EkM{zB<˙ik; JD;;3!4 2Y.$Dwiu|+lO:k$]ԜYLUҞ6EmH>azʳ/A+ԀZk"f`.,ל{=wh|_qYj5M{K$gv>cDp"'0޽5xCNQ1G2})*'>fC۝'*)"5.E2IeD 2.ZdrN6Uœ=n8D-9޵JKw5ُJ,􋃓ZUꋼ0b1f87GՂ 1t_o}{Mr7KO0Ao-Y*Is\S:JzA(:i!eҎ\,f+,Ąt78~ڋ~?[F^.A'!,iGow3{'YToҝf5ޓ[he>=7S8DGZ@-#]f:Tm?L{F-8G#%.fM8Y='gیl0HڜHLK'Cw#)krWIk<1څ 9abHl:b3LjOq͂Ӥ=u8#E2;|z꽐vɀi^lUt␚ɓW%OVc8|*yI0U=nFGA`IC8p+C:!}Nh,mn>_MGiq'N~|z`|mu}r:"KiyGҪ$& hw#4qn?ܶХfm_Ov^ܶ[6j3ZN9t9ZMMM)I[Rχ/C|W䳮yI3MڼH9iEG&V 'x`u.̀ab7V<*EzfH{]:*6M x-v쳎M'.hO3p-IGh ܆hR ]zi2hB9'S_;I/d0oIU:m/~[*K1QA="D:V&f:{7N>^uU` c/X)mS5KC߄":{H)"%,!3w{"ZWÂk>/F?RJ>FIY*%5Hg}3Ď89؟N/pgÞ tJXB-Gjsٶ 3Gzp؍H|*cyp@\첹,[up`uV,\KCB\qGiW痃[?i?S{eϻl71X:݌>EEly(*SHN:ӫOq{{L$?Q{϶(F_Ej>3mqfΤP-j)H˧&8?a?2xĐ+EV؍x0bv6 fd1^ 2ӎԥ sZR cgu/bn/34'h9Dݥ:U:vV[ 'Mȥ@ەX㧿-p0?Q6 y2XN2_h~Cֆ֙82)=Ȓ7D- V)T? O/VFeUk'7KIT, WeՔ}-66V؅ʹ;T$pZ#@L; ?0]"2v[hׂ'cJ6H4bs+3(@z$.K!#Šj2ݢxK-di +9Hᇷ絻+ O.i2.I+69EVyw8//|~<ëng)P<xͯ~? fp,CǴ_BjDN^5)s('cBh+6ez0)_~zJz"ё`Z&Z![0rGBK 5G~<:H~W>;ٍVnSt%_!BZMMeccBҎÒJH+"ūyR}X~juPp- j\hЪQxchKaS,xS"cV8i8'-sOKB<չw"|{/MC8&%Og3E#O%`N)p#4YUh^ ɨڻ#Ch@(R &Z+<3ݰb/St=&yo|BL,1+t C<ˉvRfQ*e"T:*Dᰤ*~IClz^F6!ܠqK3%$E)~?wy,u'u() C>Gn} t]2_}!1NodI_Bǂ/^8\3m!'(Ֆ5Q&xo 8;'Jbo&XL_ʣ^^"Lq2E3,v1ɢu^}G7Z/qC^'+HDy=\]?d|9i,p?߼=\Ce"|Rݷ Q+=zxB.^Bld.HSntºB4~4]%.i|҂"? ~#ݤ[tfv3Ytck0O ͧ gP\|bЯ݃5H+޹na4p9/B@Dvܫs;/f֚Znϻ-|X!lk҃=pnUגZ6p| G;;74^l{Pclwů Հ}xcSu)6fbM/R(*ȴd.^Qw %"=nluOeH=t) Hİd/D!-Ɩ:;v8`vU~Ʉ!hX #'$2j1ܒZ˜bK@*`*#QA 9WykGk,8}B6{/) ݆Y~ 1;;|,ۇ=sxy+@{l/*+E2}`pNU`ZS̯窜qN8V ['4d!FmaX-6 y:1V(!L7,RPEd;)QϢ +RlWDžuF7LFֆoM~ar*EtIbW>jqour?qzJJaQ#-n`/$fhnqgTĔO5 ꐌSYXzv9[ezksA`<dkON৯s|&*pNaJه5B5H:W2% `6MRR'xZtfC$1aH_dx$1'/v^ZZ4`9);q`F"d1v>ժbLGd~MP%m x52LMF9 E"A,S Vo}\"X.2< 5FB΢u.`aJ#Tk’"D#cuCXȉ4 ՖK(KP|dZ1&8{9rLnMRф%V Ng2K|`ot.GSGd oE'!B'Nb1{8LW^9KbN;sö!`0ݘ/l+1L#B8U֕&*?V6N{դ}Y(INBKhx2 *MOenT.a~.E jG)j{=u^K+Ȫcv/w#MivX :)ǪCZUnAS`SK6OSxa3 W; K>窜̀'n 3u0?K@BS %fee}i]>̤+*l:\歶!IZ5>H;0)N.w7ߍ|+qUߤ^oå~4en\.cY[s'wSSۘf ?.D s}Y~/J[}jX^ޗ_-/̍ݥ*n./cus}]\>\\^'W_nAqC_oO-S_sOq?B}mmK2/@DJt}=xL@5MG0ZY,\S Eb uw:YɊ|ZԘ8'ˠ*>q/E b\ R%.aS qY>W Rlz!>Z.|<VD h5^6eM>y̆@ x>Lh!*<-lo_V684A飑i2#@+j3l૎S1@:G|gRcƈ?H(m>LC,HI~'.Op% ' c*Dp*cj|>z G` |]e*:nq!`{ qBAgPSO}E`́JPu#]' 3N+;fwt[wL X1!;W$*죓Ha-s>Vzk[~S_vD.yΕ`h9U|A܌ЃECTC Tnpצho!=V qy)U cigs^>sgv"4N9W_iI NRCǔd X1Lb.u@`X]nl}!:ViI[/SE un޷(ȊD0M^`MDN74Т C>F-}$A:XBgJWq&4ۓflq6TX)ى?Nwg>]dt*?Ű~{N_w7p682~ =WBX"XA:#u-9`x 92$4_>9WvTIj`+C2"s%DƖ|2H\2+AaTaBˮ}L@dr_Wfc>IdA Od[jlec=XJ|&+-T1m8NP$%s,ig\Z:h Ћ߉n!r}_\ \5 6 d#=&X^-kOwĝJO\Vj; )!eoB4F\jtctUb.L[3M8V|&jZz/@7aV),A[5TpUZL_?CU0E [%W%vl x٘3܎y,< )i7 Ո: tC`\?c%v7\Ct!$9iç$><+c~݊lz1H[E'2/clQ.I`AWOlw&5fH n`gMytdx)lwAK~GgbJI-tq5/i ?WǠr^C/1NEU<=co(k0Q~wˌ\g,\ rf\PUH,L#L7E"`0dq@zn~+CX|,l_B'9Dcuu|~z+G q|-bb^HcUha9ce1P[;qsA.Ǎ-]W‹y?ڕ^Pm:>I+Ȧ6' ,}U=̀*Eg.6_~OJ/8V ?ç&+|t><,BLqL򱷬dS{X6"X#-^䀕#{К4i̎'QIc(<ǩJi lc*n;YKO?mW#_Η]Nvۤ9wYH3u[uq}!%9qjFbṙp IVAr!-h(@˸lɆdY[c)JYUIe:ޕ>cU+aXu U!J)JwdYq%oWM96XK"ɯFk=heДU>onue۳Ky/yH"cUMg~j2O`R<**2jw;V3X2,?XU᪙ખZ:דAAdt dU ƧBOy)3陎4cZT `Z8 gB^Or;GI~1V s.)]R6btGlɮVGwƗFhK.3}| rr /<,\Ӱ#鱟7n0?dr iG 4h:X"kZ PsZ ;FL;m yYϘj, =|9] z?/WwQ FM6 rE?I7 ^uUb_2o/Ho~$MK ş ' .R CI%C|JqKׇ4t!tG;6P! љAhjp0 C7x$ZRku I/ sB~gөΙnL-`6kW%GE34,Q F1ƀIߴ<>2m5-,@WFlOҼ(pix2QgRp":lMU&yȓYg-Ȭ(yQ<}4|03y)Ç`Y-[}d%|>Ѵkִ7N&3Q1(DHM,)^KjE\M ^TI@|r2]|R/O.&U!:_UG `NN,꺨^jV꼚c- u^yHu6rXL ZxSAZiIM4e}KJo -#tr  ~yRx&rZYƒ?Hȣ/٘P7>7u t6 d)%h NQč G4gi1O|%g\i [z6ܹe^sIx6G-K2i2\Y'^Oe}-H85? ˢA0|nvA,KXIBTsTU%ODR3]~,cT4 6Age~m@I( KkNR0ŋ`lo`0->!s<'D`S<'Awҩ j|}{!b$ipB{ѻqxwKyIø(s* ^ 4x)gyхr1:C/Cdn?Dm'XM0HUE\A}ӧvE;G.6X5@ xA"Kԋ χIWkc}D jÛ[P'=~#Jن8W0X4ۻZMͳx:!{ <|9X$' ᒠ4t(.ܻ'zA' ~ɔg9h8ɀ Z,AWZkncy Eƒ-ѻLBe7^O4 >mۣPuCA-I )&x$4 ]CW*͛Jã{h:]ŢH $-K i\:Q"N1A̴ ,|/IEm@k` ee_>~IhwJ.LVY=&"aO bb͊2g`fXԝJX1t: cT5|Ո$YkE泛gD4*+;muhҞc&ژx{56"Z(Hy&* EWC,b9L!i:U_V*LeoB質djLE\5 5ֆrGeWM',L)> `;ŴVHURϾ,tXt2q{ʕH .r)l)stҭ`E(Q67~ɁgaG" {C@įX@MhOeUF 3eRSF< @l>Uh0O_n q6&J+"aWM8*k. 8$ϥm]E9|vmMDq e;kWk!ЄCzM &xOsY B߈8؂y! 0={#!`2,6d lBA)+zbM}-; Ry)<X쑐ozJ{ՠ Oz4toCbIy k$jl: S*׆!3M)YwS3SAٴA0)t25 &|bY0,]"dJk[^4]wCk*_Eq~$SZ䒉 u)[1UMhDv6$yTr뚱L+Ke_mрY+ WțzaʱDY؀Bp$;G]Z|tV98\`oGA$=0fALg. *u@"*T4DγHWx+ Q~CXg*O]ҖrVaaκ zHI:yOUʑ)x ܡ;=;+Y,d*˛]}яW־ bm|GuʺN;ˎtuK,K󴪉[ψvz9\yzpXtWjj<:/$_ Q9XrKӒDHc ~&+Z2I4Dq\#wYg͵UoWf[fh|>9u-L̳v6C|Fۑsq/_ܝ XiL?yN@w %fҒyK P+MwjE(0lX@~|& sdwv0W qwollgZ2昻/zy =cd;UQ: r=tclwzı'v[ ק2Bsz3_^` E_4DlW2kx;q?Ms;17ݺoa>٭w!mMX G!y{) NǟT6,K;ҤػҁҚQ -ٰ] }R717!?{6 !v6zc1LޤdIN4%d'{)#*m"sH祗+қD=|}GQ&WQcuQDu~Mt|INtk0:{ҟ<ׅc1<@p+[Sٙ ֪yI-$>O2^|>E[@}^ͼbG3.C:GU](BC'bսMj'^|^єZ׬/X$h:X.\*~ӣ~w\zvt_r]łQ; 4­ c]eA ~(۩8a@] LPn d{c1 3GhVc%UG.b`2mt1F`f.r` Y[y4NۣmB!)![d\\žpL{|<v0 eaIJ5Ӱ"jlH|@f%PjZduI2#Cn_B$(ы K >de%ńX]35{7{0JOӉ7ڶZ[e5S„W^s뼚ZWٴ Y,3tڶ@Ͼ皭\VWdX4&xJg}g\V"9V* tT!Q+_ɂ՚pUW^W$z/3{{5j5٨njFpa!b[9`mݨjykRAժZ>TA)lgqgJYJm $BE#{,җsGf|< iq]; yR'/£9 f>9FGxؖwW,n^:OYx:% ONIQޢx u07h][G7țOOQީ{=p4P<[~}6Ы/w qk}X%]\z ւ"|Rף~l# }&'z8ɹ(ÑKp@yA$v{P鹜&6xz#9%ҞOyrZa E*c̞-}A$ϯa䨤 }`="DU}@kbllF}K5FG][pS(h^6zt\` 8=r[B¼eSL,ۤ5p'5[Dp] 9_Ѝ U}O U*Bh,[Stxhqm ˁإh>0GYaI0!מwiin!o<ͷ>za Xӳ+ۇv"9T:7T@#d{zfO{a $y_L 8Jh;Iʺj躑WpM'-ȧZ+&-N!Aq9WGmEᘆ:I]y ж9y誸A,(#X[PJ(ہP;q= e֥l>e{t*{d}ԯSӵ۽;wG*ߓRbPbwB{*v ^%ށP{wBjI*:gs> FGYijʟAK$`ZfUJ~yy{Nl(s(yà\`P?ޝ,G/˅aY6V\q* ePEjh>mig/?I4weVeKVtAcVn-hqZQ`5 QHi+!]]td;a|^4-Ĥ32a(LI}?ցU'T:$Uq_).xBix4>k+߾==Gxi =RR @ %eEd?(p.â/2e@Z΀yjHG| }\7`V(>w^s2eb?Nz(5-YjVx:8h/^=tQCO37hĕ cY2O K/,Ҫzph<$s3[QZ-.8Y44xe׃-2,}U d"@[G޽Yxx&R-D#;oeAhu?S)RŰz5T)hdḶޭAz [鴞5]Te $Mw!.Lpչ A:^㗒g‚:kՍULY,θQ$L)F`v[t:DZRT7Lp9뼰W{e!V%hvu鷩PʅR][;Ceβpкft1(GMT~L:"vͼ,#VvҐ-1"׬_Gu:#;yâk}qr*HxD*ɻq7ƸP:T8*zH(kR_<K'?.rX 8 Q]g/7j.+|d_}`XC[s!OˁzH̀97:QjrT}e\Rgݘܪ C+.^m*Y_ӳUTTvlD'h-\~_AMrD]5#D]*i l8&Ω:8rJ-u{>AmӣZ+exs:9qZ4R$r>h_aC[q?4tEji߄%+a[,,Rz&1!ViȘ4C6SY!&[J%P٧aTI!ӓV 7.k]JWd .d d9$ԙCRޓQzEZ<n~x з|.p_T]&rLJ ,2^SkGhIўX6kTK9N~@ba e(Fp'|VYṵA,̨s4)FzR^,U3/T^e?^C <:ueMWhIN뷹- pqvꔤ%Q-r5 mµ74jJna١POXpFS+a\%P(Wk#E0E}O6;"!9rdz3v˙ۮ3iOKi[U[Q-rΚ&k/:aXB;mAbzT>XjpXpɶNQ2$s:ˡvizjfha5zb2YL*, d֕XFl6k,iaxL{i4Dz]R,=Lb+dì oqDcNj|O4һ_yg7eL9$UBDzht {ZȆIfۼ{+76<<M#+(G0DiiM3U2_NkfCqR8ݯY<9X `> V7.+KfWWUp( `-.,L粖P2E};.&Ȼ޶W~`AyNJ(RCR- =%rV~GxΧꮯ~gO}z >X{.?~Z~s?{'st q]}/-w}ףtDxrjҼ=Qw0IV/6g*;n=hõ#"|ڔh::Ͱ+1ߎ+Ej&1Q1aYbKaDᷓc1bނ0?lk#j™r(vZ4WiP[le*ˌ͵9pte30bX~̶piiڧȼrg0|D|HboY8:&0Q ]kޜ{{L-!@I1>" ΍uᔷQʔ ,W& ÿ#PޏΤli2^a}I01u=hݴtx.%,pT|6[V(;I6DN$88l 3qɱR.e G 7]1űHɬw|&@O#qDHm# $L$\!eu`~CXUNA*i`,Y+)x--&HbM0HhD{32)&LYmC|#mnVGBC[V|Xѱ=%qd)f=Ƽ(&WiB.) a4(51Ť9 º `$31WتჁlXwDnA}Xim"!i6b,{C̅9TLlA*&^9YKY[ƄK:x&%k!R>K 1칦O6,OC Ŏ:ŤI_n< &,NlQ '.YiQ3S&7 IJyz"ɕ?X&q/Ky]3zn)FH"$Y70H cZ[8td8T\Q%%a^Ѷ^z^5I|*½e^hM0Um!Đܭ߱C܅OS<'AҊ%Zb'AMmI4/-hB0'#(JbEjLG>cs.$چexRAS;5;$Cfp"1Sc896`,y0`th!ZpTC#&^nIp5ZcbJ+zp[ULsXq#9@y8Ӷ{8"olYIt;3 [Il *,Ff{mt,`?F{9aMtVk%'fQfbe4vd!>$z(N3Nqc^ Eզeb ӏ"v{tiO/F[r[(, UY=`l,2FR|~vf$7\s>if}+PJ[,4&Ū& H",/*J}Qj?npGC~ 48Ze+G@gI(ؠEF"1q U~rt][5)}Tlіpեjw$A0 n&1Zƒc- V[F֞c$ Sڔa7@8RKG<tatsK ; 拒-hy2jA6rQ5oS=08 ch.,fp܌mQ2 #2"=EJ΍LhEyI$8ZIQ?jLwFTˑ3Ҙ-QJ;I!z9:{Cmx/+)Yug I$ ^HpTj+Om7DlkIB0rэ^h5p V PkYhƶ%l!iHV_sB?$&-s'NBY }Aq!\|uZ]rpT)%AXbB*s4Td4Felڭ aFib2 &?w4.OpO :n*}KDl uT^tP~QobJOtLsnr>>VDǼ5x}ƦA.LM$8sLFbkk_Bľ'^g*Jd9.y!tIz®\/퐘G1Mq|p ř3fzw O,$o5 h$z8t_f>g$Hm&nRAWyz h5(p~!F{<֫͌VO4' s8Ր$ŕm ^5rK/a%<_!d0PKD (܇772Q>:pbIpL`.`o|X^o`I`TJGMY`Q̫<ZBb@t,)@4upiA+9*{~7*_V`2I$KvpKjJ# G̛%-xjZT! BkG48ŋ%~+k!yD_5 &G:0}cꀌ<.9"VjpH-UלR掞sFGyh[Z~+fejYe:#ہŋ4a C`s'M|a&]tBjCNуjP6;ZdI3Y=?TO`W8#rJ g4˾AꈜTtqey& D}5~p{05E`ÓquE/A֣1Ѩ':5H<e48a_pEds3lvenAElS@7w;z򤀨 vPVڽa)#na@ lrW+op~i;rki_m7u?OYxX@W\ (5%G^=Z=@KE?0^)>W!kO֏7?gѴs662U JH:V ,-B 0V|-6}y}B@s7,t#Ε E&xCV^`$ (7Q\|o-4F:+MjvԝeU:xPI})c-|m_EY='j4cKr`Q9pZZKe&js)Z\~ڛtgһ 5fQtTtQ  /7ΐh/8Sw) QFjioԃуybI>|q^8"uQsh=xgD.(T/-8gdBv%%uϢ>m)qImj KV6?D;pxL6doIB4Y.jhs>!qJV~VJKTOI`c@%ٝ]{!1&uBke-EI$r/=.QJ!iL~U[6ᄘ2«ЊE!>s"G e2YS@\gTs'0݀u]itCJ[L):N<ࢗF_u;Lu23LJeLfa:3|Q}C"Mؚ3gO]Tc$?Z]Ƙ-*# `BʸN-tIe!f0..Fwb(*v.܀*J*il#$wbL fQ45R5IsWϭ_= bTN?C7x=PaEδ̬0,t|O#gCM>)ͽ=FpuY dI'A8j*Ѿ+9MLK?+ߴƊvUK;dpVmkƞMQ?-UzftHU!Bx ZY 4NI-96N[1c-0cnIK 4(ETs)/ЯI`NKL :E%KB0B74lDw@)KEtKZ2V' ][`| Fh3Kܟy1+`L](e)#Oq|(:ʑ$&ƺ\8a _Cx>;¨hUNk>Q!4lQHJљp#`Z4tvJ`u_]V5p!HFN{ZOO{ɷ{_}rCΨ#Jso2V SfR"ۿuO{Y80A—" _o[Mzx1c|61S;ki> Ylxh_VoXau=KTXc4CEgH)ʭojVԷrB`j\e= du_E%+%& L͈ZrK3j#L&>c،/Uոj3&7ҿ>g2b<2#eF2vVe KC.-[ ojY5yB-!́qa~}ݵX?20,mؠ5'O$ޭl% nFe]%;҇ɜ?137۠uc'MɤŐAI.h)R/\?f3uZ1 >4]t -Z~MNF 臓7ˍP2kM3F؝`5Phj }SaǦ/mF[m 4`g%O2ߌ#+SˑoƊYY'@my3' p-tfm)璇o@݁%ݛqɰ7cVT5~Omq)q6mOл!fgݨ\PAC|Ie_S1f8tIx$Oz36IX@wOd'{Tv zp?<Y3nkv7X ?$*#a~S6́gbzӞ$8٦պ*ng_cG7}q KMcҀL|ƍ&䱏|iF.}QF35#)Re !ю^q8fvȞ&OµdRZr}GԲ5Lk^;whdBr R~z~|1q}D%N;(uWO"} K/BɉB!!J3)Nih/Q*+0w1=GqƉGYm'b ܹ,M fqثIͳ֓_2ZdrJ뒡N0Tf9Tk b,8N3T Ǭ"$3$ڒvVvKTd뒡i kjCAs#3HxqFXʘ6%Xddd(Hƪy]2 ޲ zC;p~k3L/TWR~ZIv":(I5ڱ]&<*&@;=x܍[A^;L 9 F|M^S̚"PJ_ q޽ ֍~#[\\ҭhhߏ 3!.;/+lY +Ans%1GCޡ ?/ a>;}(8cft GU\}P.D~ӇD߃Sx@@SnTh'.Fx S`k^džq?"4=Zn f<2߉Dѻ6ۚ4c>8#?datD ըnU ʗmVmLAQ~zZL=yHp`[ߌGݒqT򇷃WPbcqxCT1 ik ?0걛r7#A߲5шZB&p@Z|[E`U//oo3sgpʁ1o\O?z&l$/F,2Au\S+~KLRuOëkPd`F`lT.UF$ WO5@[rjbsu+l}/o{6JtWYEG+Sa/Vg_g8;GVpϨrܩ)5X#n0 p 3R%^%Z=&ʢy(~n|=E@H!OnS ROQAPmʴKrXԌ9ɐKDP*"F`)C'Ls'- GwMnmxcYFs+4n#.w&SyT0EDuGSQ; ^Yb+υL)IC( faWgYj>iې yG ttk<(h*WWXNi>iT5DvHUsqUC=h×-Wí!šČquӌ/M |FpXkL`,bbl+ʥ]ЯJg|VֲZ6YlJuHYÝY aZE0H:I2,K-<^<\zX*S+3pMH&t *MqNz41,3ȧԙT]hoNp9P_Z;D R PcsC:t "cY*`#&sI&%ɐ*i(ꭴ3)``E-2 8V4fyˍ.#eO[J{iid`c F董4&"Z;nҋ5 T1FP_*uz@ddݺ , S]/zS`꨷f-?{3[51+r{lRkH^Ç:Pp":mp?ЀLΖ֝FcةfrN«rZ{55PKUT~;4Kմ JMv%6/QkEA Udb/V ׮?$4 R)wh a0U?2tfV3I!Y*IiJ.Tj%h&n3` sT B_~$Tg%7or iLs@d= iP4fዬϙ~^HHA0]ZnXǾ= ݸ! cB0r~Ȋ`[Yd_D]"L52\65 F&dB6D7Psi̡]9Lc5VfIh`$BA6JF \jS6R* Вv Tsf4J\IEPO]ޫur]+2\Gii/ Z妧ENG~μ}Zy=.[Z$e0? ] „:ϵ.p_c~7ڱe6<{Qo oyGMtrƓ祟<9'(tV ѳ(r[wWgG"U }kgX_4cQ|cdsuke%Us_`.nОC%4Gen(nޣ^^]%Ov̚,p˺pKt )*On3~nпߜ B{lޜous evf6.ҋVrenh*ɷ7?P1s,Q3aJM3O <='6μƟ^FR|H:-j'5TڹB=FzKo,[zE"עTQL(+u 3]7纍{Qԡ"Uַ_d2 *뎗ʬ9{;s3Z #wa[3vʾG=dZkw\C؛\G91tnyym͋]37C6Wnn^q ωݨbbvSnh1J^Q-Vݶ;=ڨz^g3M&\΅fC$KUvxcGߪn}5[[yT:w]3"\s}"/ej+2>PNߧz\_/Wnyi'v_1뼺s7Y㝏xk& "dxſZ܄7߽}5#]o9si*qhȉgq7D7;˳ӏ_V۶Q0c(y>2TBόiRvf%IߩUru;׃dfVns\u{{ɽNl(+UǗSnK1\tT%-L&x\K(I Mɠoc'8_weoӵ_[kȉ쐦w(ݟOP$LWݶ(y9F^)x3>~:+9aGGW.zM8J̜8:!z'^2 ?]nQX˘sf5~th#g_ݣEg)#47ؗ:,@fH!0%* Ʈo׎=R u:nhaWLҊjMcӄ0Jt!&eQl]WA @_W+h?WFV}57Υ\p[6$cтah~ #phGS yZ0` v(\Ib6hݠmN T́6,EpB5Nw3L3< c0ϳ T, .|>Q3<Ϡ̭cgSisO/Z*\BefeR½ͦTQ}mK6/=RLa<l{0P06 T+2t} Ύm09P / 0c3ʸMAhiu$BD= y@ӭT͓}:b:]_eYHT&gex:p+4<a"ڤ$S1TA!Dr)%\qhN9Vlg^4ٳX`jRl^u^;se+?5"ǥԹRH[W>YYSm@kY4@"nVxn<<2,2RkX6v)JĸPM<>r׍֢  4c]FL!Ӵ R!-nL"*$L:EP-; F)y0 |[΂+L\gfTжf <[.Arl mXT]TlqwpT:(ݸ7R4@մ6?rmޚ@;޽QRuSP=4[5Zu =+#CaX,G~mx2 |eYn ˄"US8z@"hSiڐECE+^eS+/x]/Cf‚xbi CԢ!-Jx9$2]FA3YW7][]GuYdetL/d!4چ$j \B A *A[G}53!Y<@76}K6k p9[͡ 6hXGJ>j ʿo#dтDBB`J4lUq*%aX!+_´%xX4YHmt*kxdHeZt*ZeQ1$/Pueд,V~`DDpL=KBsh'7jZ E8_''+d.珔e ݸE S@.8b]c3|aԕRCv@=P[srǿwݍCR'~țR΋_jbK uIT{H 9jb*mۋ7ݍFcW|:\|,\[q@3dޢRmNn.u,P=ld]L٩uk^h"dϹ&`qR?p݀\9g8j6P B~\rqxܖ]谮% AxJԮ^jwZ lNUX=AfS,HMjCM\^5b54ڪ~Ua"ڊт㽶 /Z48vVEhZ0c׽gk7 9!BfHdHud6#,kc]#i#5n#fZk%ګr>ZAsQm+'g3<)rm)?@uf‹+/۳ch3>yz (/Rg)h M׍ϵr^?bP7}b G;L! s<||8lc;v[M?zԇS_Uׁ_hv0IQ^ǺR;,6Tjo<wnpMozh_bOwoSGvRHݤ5Qv.Fu|o+jχwlŷ`;Gdw_noך];].Ow s=Q;뻂۟kOyğ~6HE'P*<_za*#n:"T@r/K"= .~z綐zpb5Y?W/9\6R$5m[4Ivm$I幂<6r Œ7̓㒑r䨀'>h6ThQ|3Ÿ`1꥽?uǴZsI ON~߫ya/jTI=,c!9fj0RU*:wsPke+oAe * c#;%*UAyl=Im B5<`34X:xF,z/zLd B5= 4TR9DjVkTv |Tuoou{{>noiMT;nou{5xip )sm[޳r{!%|TEtR%t[st{ S'"HA24V^s rX[}޳yyP*;z*PT"|^j͟p__KjnW/W]  Z9c`30׶{Ij=XXQ;M3:OX|W݉>zz\5(aP_{cdvqenW޳2pux0z8=RIu/r鱋۱ &nᦏO=M)Uqߣt=]l0YH,l:jmE 0`~2q_Pr@lPqB@5|~EQ~FqAر 7T? I+kZ>CsxlYA[s)jaAOyH)a%0 v37M\_A\bH-ćV0@ػszй6) BbDl- #pH Mľm;r2M? j:=V_-__ 抰)}SEe Je# ,Nc#摀qG:]o#ݫǷD1I F8"0R(՛jokg6ZV9a6Ѡ8%'E#<"hrGҾA˖A2589sL'YL5nJŢr@a]E<2"E,#9shnEBC*fb8ˆ # 1P9)>7сMeo ۗbtn!DP΢ՈXK޻U|ڬ:%khb8?c8Z6RY/w5 ƺ:XOKYA'?Z#dC]70x?yI_xiaZ^Tg)GA#k# [pumB"NHv-(&5Qh k-k$aއ a2R K4I%҆,Y@% F :OÒe#ŰP]- _`Ivh Wp×xYOZJpuG.[a\M - bFcki4ctT&y^3$@g @E: y Y[4R 9.*1SlzۈrOoo d2 R,H9|I-*_Kӓ=_\ڶ9_؁Ad)/%3|YcT/7C.>c[vZj *Hz3haکM=N*힝vOfn6TV p vP3q.7F Py7B Q3FƗ7Hy}SoLx.2a;H+1NTkHYOb`J)x[|n>eUqnWԎ>=`a,,(1䴔ձmFSnka3IcJOcFlYZ4R~V[a :Lafآ!N؂N> PF Y@q|; m*sە|q')-y9Ϝe8y$Ġ۱晟gov$2a g`H9 P]M<\85M&GaGj G!fcH1` ^`k-|F2GL:Py2_2Rő@H$5pM$|"!mDBdksfςòr@80 '{[@[ LF"q]f^Ԛ1:z?bh4Hj:QDgdhsfHAǗqj3+,gyFGhNȤ)#0Pm],}q; 8vU BʩQB䠤.)lc !| 1<{B;ptBO.8kr0e^bjSbB'ڒmqtcڛDO-D#H)@GPgfg3]Y[=4R3HV]Jhh8_1AYqZh-ѽs|wc 8n=H1HzA%U+4 u- (w4 )dAkH923CTuuON?YĻWҩSmC@YF0a'{i 'fmxig7o_:hSpZ֒biARkZc+8.PiD !?=O RjDA2P7ͧsT\c]t6 H>qDT>M|DJeW?d}rkyZ9^l[%4KN lZ{j}JyaR.m?5u'ErD.(`Z|ՌK V`^=ae!+,*!y ;tO4胓t}3#dN  w3, dS0,,)(T],PFDu C[H!u0%NI3#S` w=aU70Rp@5GFʑpK~ 68oᯫs }} E+scdŕ%`%bvmWV2Kz~p A @d՘C0R/wQl)<(-%mDBe^Qr+p[.Ժ¼w[7nlP^IZ`q0]uA gycCƫX0RWXŦzÀdc\ɮߦ^`@2) bN Ob-d+hGtvfVC@I>U#h =S fx!0,,)*rgߏBe*>Kٮ(`Rkeh ˵YFʑq TmlNAh mbQb-pomlo-_oݟ?{6l@3&`vgY,pA$YR2 V7)zbKi)tu׬blT) Rg{kb?D% .' jgyқpq<û8o&dm@_n%Iv&zX.bLjrٻf8w4evR%cLMA{5x6]^r5ڎbnlvY)pum?kGQNaxHNݥ)IYCVf2 6HYWSyREch%Ү:Y-C1* [Zpe-<*>Cr -iSm3%jqDpknjhc൶Ѫozp%y' ;䦳>Ol#'/Pl{ ])(b^)w֧qѩ}C2pӓ[z\/b"*)tŀCvQ)'qGe˨A@IXLHHH#4K"M68M$Sl^Y<  I I)b h=6X߭`/Ut'; =0GEg`-~5(Cj&v/k?<~,d|h_M5q? >L`\>[xFm9XjpvȀϪۀyybyu:CD %H nkMR nтۅalaM/R3RlC)2' 7C{$޺؂+5o%K ϐ䄦K>шQ-yP-~ri3od*u^ʪ6ǦGp(Q[qJbMJ.T-Yb>=kٶ8~ tjv6JDT >YC)FYjR$F NRōM ~=/$H/yY]tyPp'baNgˋ$XHjb3)޼m-=,yS=K \Bb%"= BS?F GoW|2s&mkagwE%,$! !JRC>)B$;U(0kH5v]#}k2&{ F:j>G_"P"iH.۸Ը7VkqƶnlƶnvcgfTSf% vn?WY\R+i&6'ɹ;Ty;8 FנV̸;oݭX<`: w_dIAͻ9| Px6 G/(uģ՗D3^N-m@l<aeF'H4x&# >*ˊ7.< /*T/ޮnj,~>FNJQzOv˕-qj;7Z{y iF2 S`W2bEĀhZC}ۉ!-00`n]N%LAW=W"|S,:v?Ft˞&3 e“w5/`Ǔt0-wpr;/l-odz/ ?ʹtg3d0LNKM<܏^G}!?/ /C~7f~2o,$fL0p_OGﻷCNTƓ8saB;SkoT=( O,9XO#nodw$2ؚSJ#LL)zF$klHXjs!a鎫\z~xKz.*IO`f,IY&`ğ]& p3QiX6"^CP= uݪ[3`tc&,\dTv>I>݌ӼwWEŐ>*Իf{tᵇ4Ha}TnG_wR9eFŝծz 4j~ti s0e^\y E"1ژ3IфLg ̊D,)Nџ$!?};D|Jf9&7Mρwa9 DQz@+}APSOb}nn3 !֜= h4ƺ26>Z08įQ7z,>XXerbVěD4mVM_9 $͟Cz@DAA ' 0!0q&ba& 0!hxv0CGf5zGU~2|rY;c/urY'`笓_r2Т͚ȫVnS&&rd|{1z>βC|G.$yq+b "ݒAqo5[VNs۷;w~[Z/u? b0q߻'sh9{wWKXv*;Y?_d9lz3QQX%\GUh;9DW/zO¾c+#/b\Xۻgia>{nq5ጬ0^0ߗyGGZv+47~~= lx1Xh C9( j%[V,1ΊILT @@_dp>Ίy!P5Rj!Ɖ*3a1qf€ 3c$NTR`oG!ܲ uӭ]@[rJN0]p&]fpS0+WGw ۜ FkJ UcybS)?ܣvݞ{o5B$,=G %@/uyT `g?Ly?Wgwq&i &n48N-K0cqcdf%TP KWfLJ"IUS [AkNҮRtVU)t'IY׎|v}m&>7m>!IapC @ eXkJCbQC<>?<.E @+Ը!+TI@Ooפ! h%҆I3.Y BKz!l5G7JkĘr5WM<_?bPS$OޞH5RȞ@dz6Y{"krh $<(cݛ{p-' 3t2R\Ξ ?)pe kQCtnxjbR5Snª!S#X5\%nxZ*(L,tBjRM9~0ݚ*k|MAp CF9fyZuPRAک>Yt+8mE0Fi+-em8<+C;܇Ys$Ѩs6ٌ:NqeYP@]"=YUt~mgxBT[IW2x9knjmfA =pCe `|0cpUah 9F )!g.OSsH){{yͽ>ስ{]ymm21Jl|FN%BX(@zB9P2SE[ cJw~+|b?Jʪ9#+gQw^3d]$}3-_;&ײW6_o7y85wZ#j&?rM k#q7L't=bsX{$mj&@U  kfrm*nx͜JNϋX/]bҝR\Cb.Zd!ɋ\:y,˚s'6=[C4fiи v RգAidrkZwxj2V=+lkޙݤf`Nit<4\40otcA0\%4h9F5`v&^)0pc{x@R܎U$o$f.^srf )ٲzclTj՚  s"J)RχxnՍѪFx 8e$J"Pl r x:pk4-J΢fjCOf2Bp "9'9+o}bF|Ⓧ <%X%)@"QU,Z2%ZR഑TH@ ^OZ.fso]\I"\ z-6JZah1KNEJb V`&Yt3(!Q!c͕Rn6*:˄Wh\ vX$)K.(<*Ή<Vó4 <(EJp)IJJ#\,^CAQ**u Lড়2Eb (_XqVQLփ@z;-50!Ю`YcH҅PAޠ|ͩ~0!W0c²޲G 0S ) ",C" 4jB٨ݚqHx{b"Ȥ Ä| H DUTzB`1y*"TU9mkL64-/ YXB{S2ÙX)0J8 LseԸ`҃M:w@<@7/:* qJ u% 9.Qud$E\}]sppP,f^>ՏL%PH/%Fiy/Ed$5SH@ $DO];2l$CB+|΋$MOd38τA787vhF1!.UQ1kP,8_W)*SNN0BN#Sb0`w?ܶpXAR2f=fcFw!45MY 3 ^#x:py@9>m9@~W2z(9P4-Z.]UcNP~ hR q!B*|"2D&jZ!T^1 % Պ.KhꘔC`lRxIBqIJS \Pk\U~D +yGŋCFHA(e`j?z~nZ.łUM;xb'!v'5MĸoiŜ8*YN1D*=zrrv'jK)+˜?ހ/e_kMϻNw;]zX~zw 6ۙ=ܣw-\M[mܩ?}@pqVMz{Rw?7}\}Rt`7'.V1hEzP>hs=> 8Hfafafafafafafafafafafafafafafafafafafafafafafa3)3'Sp1x}7%0ZoᵛE0\6&~&#bzy59@a%1Y@K]buq ,Zbsqq+_nb4^~ń+BiC3: `q@V*kըf.`  g`QWϰ?lw ;6۞Nl0U߮/v+/V B ?+x_?|0QwBDlp5[FA?M6 /6cu`2VV[U[?<uYh˜Lb\?߃򕀅fq&` GkkN)LjȤp4+5ݜ10 p 20"sI]3V>#d `WW|\ _?E\c>[j7 }gfsu^F%hz뺹5߇/[f751EUElfB,V%ZɸHQ·!+ߐ :_/Onm7slV_>xw틭S~|~ Lw<5|Rj~X/~\.5>ɲ~ys{>D[nO_>}ad*Y.{O?y7헇|}}_Ds~>rPA XMA XÊ X˚̦Kedꕀ62%z,s ߗOy[f C}}@j˳g:st X:Ul뙀LF2Q`- X0x&`! XXL:<0TLQKt@|DYK3E'`h!5V TLXz&`Y\' X'G^x8X/r@7W6ܘ *#+j ȷjkBu\€G6 Bsu΄#/<wV\&Xp}⣂uu:fVݎLG^edcsپ;3L sYqĬ+G2}95%%2@xr{8X{ XLZp YeȊpΓ;zZedu`@Ty<gf4Jp}WCgwZ@G^q8Xs`P`6mJAV*8[qr7d^s w].1SVw{=Ex{LHU&{RQ@"G|ZbX8H_5(±y<yh>+3W\+9MP!|jD.Vtn⣖;Ek~E|W9;E1-TMJ[rMNٚJXG7~uY?/b"U%$1IU(6Æ]L\]jԜ*ܛIiGJ.\m Tf9#pAv gdoBMU$.VD'CWɦq1&4JpMVtl)km#ɮ`۪ Xx@v3X @6E=m=ۤ(RŶ, `l>]}s:jr9VЊOzh j(j:(U3ΙUYT]?IbׅPcN"$=rX B6RWD38fԬ05Ǿ}J*|M1fl>;+Ql[%l31dqiCW깦f!IdDA+a`s1h*-Cٚ|? etbR)"jb7!pU[VbUFBVk< JI n˻乮7=<+^8: @k!0ep v}ZX?j"fbBVCID-Xʂ_*DO*RR["441[YV`1rԺ S,x**3п}5TiOa 1 ݲ=Ɩq[{*PB#C!%G^Q=2LQ\`Q/#~ FcpOg2}˨Se xLsX/`t[ Xk琦 ngdfN 噴UAŒ5XbUh2 %Ndp-C%OU8ܪLz4b"ul@zc6P$V11T1oCXPI48,Eiܘ59Peh\Qs( J[2JPHލ bTuL@r抩dJx̆6 2l`7vvF9 t JP W4"Xa. W4&EA\:H'e &w͘=3Q57.WR̨Z!0c@ "5-+T^0 T:x U:? VYܣq=)/. D,/xҁD›@q+*qΎdA X?`<9QT2Q )r8%-C`3Hd~3 D…`~i3j̆bUq RLYY 6EL _"!2 .H,9vQ0[ U&y]HQvE"B Tv"#DoaL2@.uq7.RdSbc"u%pb. 7ӟj -tTNCН,S z U>_u˒ΗBs>OsE_X/KUYnYH_mNGjN{k;!%)ϷY>bWDY8Y?w}5Rh?.&y]?i.{^BqO\OO8ƛOG^ޜpa_3o< _)8[ёlroh弤/rY*F?D >*?2ܻ?{?vPxC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TC5?TCp?J`(ͳC%^ Jj~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~z~(o=:<?Zy4~(Ͽ?8PA߰j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~j~zY~k_yN n{Aٷlm0)Rj>v;[lʺC?I \NfSzԴ.2 g_+Ξ+w_u Y0ÄX, "/!},윴Y|ߗibVS->,N(SdxaUΌ6HB@0ͱīd8$`#m(aMqq,a8 Xz4ÿo{\#˹p\ X%컯B{2f"o߿q i|*w'}|~^C!( A<`ovxW9(]`m&]zPz[zJ ZyDJ o}4~k2^d^xP* LWFI `hX5+ByKW/0]Y0GXx5+Bi]KW/1]IdLZ+_(Gę=Oypvl fc@6Wc(6OXM㏯&+ܒ_q^W&/j1+'c />6a9Ii}ȟi0_Ӵ:cI NI[/fGA }(EW,;o uՅYY7Pv7~9{~+YzUCI s4/żaWti]ã/q]\(;n[d|&xiJ dizv]ݦj?Vnw?gyNI;os"iWw{vd]#aƟ TR9F|D2E\]v*1E=&t[e7)WΒG:)eT!׎ 5'=$ 9m?h g_xN^nݴg+) *-Ao6..iyEx^yˎ,;T1Mh8KB.|Ӳ^"eu!sVs&Sw'b}8kW>_Ͳ>OG[f/NvVZHlyStpgҿ<[c<йO&ӇKW4_6 N)X\"55-OS.z{WC͛q0;35zy~aҽ~LS{`ϗtqr>6Lu<2OtZ:r%7E.:6?yJǘo ~6{{\i8DNe/?+N6:OW26z3?n%N,/I߇ 2M6, W$%L6)wwͽyi2߿$~Î_6Laj3YQ6/0]o/b냛-~p-ϋiaOaqw$b\G_7Xrխϰ1~ʾЗMpCJ7.s-Dt_N8.s \kTq;Xt [_qN#UkTQFUkTQFU{T[/U{}ֵ46Tl1Er?^6…4ڝ|8{]ύaV"a#t5V{7WF[Kic<73) Gd"xL]c1uehԗhrR1W.JJvשdLXx0fuǫƸcfܪ*MYRYB+'J9mdxd:uM_%.$1OgQ_~WpޖF><.K^ STL,;|HWt.Su83@d&ȿ-KS7Oi H ᎙ nq7e2C6ch{):̟*`XH?)鳖F9Pfli3_6I"$SwE4s)J8g(駾1YY'ON}꾓( u1'g_K[[?D?뚘rg,#ծdi @iP^W',B(9ѢT0Ғ"REN# Ykc!Í6TNJVCZ5ooNH?(SiDrIֿr9Pӿ/q{;KٽEt>][2ߋòZ/~9}6`raQ%S"uNZ "( 6J+<~9 '^˧Ak6l$ѐ.<5o)4%aLK9F r򡐟u-X ^~&˥"9Hddb68o"9{D.nPt GI!`E66|OEjQoE!YbDQ =(*3bQcyB0$[_n/P>/w-܆>߯H?9QFX .BtܺB4 an[>f1bEP&XފB$XWpo)K=b|(M"S@Z^4 d[\U noE-Z3ьxht8&j\#68jPpxGTs]Xk8@iȆdz#]c-x{2G#|\ģ^RX"P Q`zP7F ox,oW a!JyAdԊ-NF˙bY;rA3fw f0+9ukC:I)k"&J\ EڄOW bc D- nk&[Q'G Oaxu5vftlS7v c˼ RI׬wxh050Jl '\\P.pB|(M`v7iL(%^I WU -f[C- lބoH~oozM]LQQ"% I@)`H !Z:.&'OHӄZ>tC)*,`<+N38 8#c;NDb6@- w;D2J.S]&Su«(1j)!_ nb%5M_my'B?yU/z8*Nx_?cx9>Ư_>,`otDkhF']^x{ K I$h^%)YjAln6C_;k޸or`7[/?h5]ν? -"[D-ʹC* RR2bG'J|03^KuڶQ 8>b‘65$Я٦j:wqۨpuzceZR^(0NDdV$QW\tac6<A#*O 8û *< .ZRRpVrE$""|(*O0Gy=i(Rz^pGUufL"2zM,OMȽi"Cu6V\jMO)̽둊$GF|DW\5nP)*}_BfC\07?Pdښ5Q(Q#A$n7%aMoZX5Z|g:4Me2DXjGAo5Ɂ4¸ Dqθ@- 5(wu5CMmz9'^/2vne>k_v .F9r<ݵ|7"Sl;*8\u4KX`hV4Ӗ"RAZtb- N-=EZ0Tf\\x2HISA3B- O뻼w}HFJcy 99 1mT|^>31o푓tp O4uɞ fL&eBEh_C!b uDiSp QGǤ__o9=a!%Ef%"J֛b)cEh_!j,jP+/mP^*X[yf+B+mD^>hF-ja\H-1W aI4N6 8DxͣgQS;v;0۪Hl:C#y|#19xGKJ#lmb:jC1tVg# fYi@-4cj`Ũu~qC9nPGlF'3_:xԜ@=ab뒼s<ⵡXal<;ڞ2n*Y:FhKnhe+d) 1+f‰(S bO[C-Ώǟml$嚒:-(_+S!*kKUI1N:x0F[ ܊ ǃM̹M0#)A6E1m\P?Œ݂| ̌hn+Mp J-gIkJmp0ƽ))OZ>h1bgǭnvyiVl2*nBA k*r@f /JM . fO\i2wI0H4Oֈ=lk:WO>0"c%Wl8tUŶ4#&E\[5bp5 U{%<.%7B-_ܢ]Cn1U]/*l:sҿ=z}aZw𨷉1@.dmPI3o]ܗm `D#&t9_S7>zGCLMqrTD`J5cFo~Ef}YNZG uviZU7 Tzs}`|srizmz|uRgZ[[ Tk*Tt*YsH ]RO3^EN>;ØCa3k1k縝˧lzE%̌fy@X +fL!W.Î>퓂#[sA(SsfFG)*W񵣴E)SQPa|ʎ86esQKTF<1 mJܒ՗:K,1rӨv:e]y1<6G XhcV-2H|qϒ|(9,i oZ])/;)3mhvBn yZ>sp 5JFnc],bbI̩uU V{)iω4dVP\Az{96 -+P˘1eF\ I6'ч@<3 %<& AL:xQ4(C!ת ڽAvۀj֣+3]yj%|OF=Q Ze -C.0yzu;ǟnY iv{2P@u7<^O,+|(#FV>}d*[OP6]0uxIw=;j>egTԲ$G>)<Ϳ!y^ʐʪ4I.t܆UB{ aNd#Cd 4*-q4ښ;j@U23+Tc \n7[~ }:]_q<>~-^)GMz pf/Q#<ґ33D$ѩPCq(G? f;qm o+pCE ۋX8Ec=ۢzdv]&7rH-Y-8듂_l6z,+ʅ 5NVXʂ6]Rߘ1nY=SPe^+ꓽ(NcG_h6F0n79bUͪUբxn=btxR wJˤXQ{l8R䂂Jt21w"jPxS]鄹D&KODpFobS2boF|Rv*W FLyD+CL!@r֜kwsz ?zlW~{یA~5M*6l8t7{82zM,TpQ S3v3PR{q=5Q}:xTG0]W0T,Zzl7n{x.;)Q+# Qt^6,t O!Bv| |la5¢B#-ezOӯbP}k8v[ڥ.VN8oGi#JwY 6zYMxFBT1& HD])kZPOlit4dsg}g>+k],b.Gbn hPAf22=|(n:OU}ec2)fdMzIryCr~B鴏UQNm9b/åu7$_qR<8r'F P OcjHrIi_cP!t/1ĴQqwtO_nlj iIZQ7YyQki1]Px~BMQqT49&ӭY##NǸ2i|{'Y6#j=N6&i Q%R&Xǔd-ѨCc[CmojF? 4V^xĤ>K哻l*֚i >;jve~W+,әnzd.5߂vDJ(JR]Jiգ2O&lm.y1=uD۔LZt!Lub)+y{z~v=?jPG쮼|\R%{mo|:6aLBN}6+)yMZNA- QXnp5^%g Sa^8n#W3"EJ]=aoMvnHu;Ka%[Rے]-ژLK,E`󺍛# Ÿfoh,_YК0NBc CB,Ղ L%?5W6Ж0Hٷy~w<{w#]/6vxG a% >@4ѥd̢a, x'g<,>gR g\8HFvd?rJ>V08B KCLܯFv9Mp&028ʔs87b~ Vrw`B_rm/ArĻPdNgYPDuw\ ,q>c*^.jsۄFQu$PkH5 a@DXg&ZB_! mDJtBT@83AC&M 1vQRWwΌT%c0"0y#+dq8Ҿ5 tet:cպ׈Ŕd h<1Q,k3fCɶG߄O1:DMmtR0V$c2a( xֶxy@ h 흟ӸN9:=(|f(!~Ӿ \kb4o4㋺զ @ b6P>qKiKP>ϏY"FocY I7Ai. i>-o)D*Ib"JsA\dDR#MMxUF+5| 7QX峞-*|%AAnBVqp KKTR^^VhOrNr}Pj[Nb4Dqp 7Kz$9f~E=|ӤY݄1LxP_\xo$867!T•QYF] 8m,&o++$^9;E 1VQq9/)hNrj3I*P:"f $Y2Nґ!E:YXY]nAGdv|g 0F x16>S'FUZ?= >C903,|g{`;5US0ClQS,<ι=Ao5wh,;|S1B BZ xL߃wWMY2SI q4u395#_л W6+x Ԅ6D(ARЄ,d=G{oq`/)I2k_}T/T湮Έ;ﯡxzSEkc`u\O1&TέA[oFqtGOM[ a"޽C7ySp#S%K8 7&T $B[آSqmq4ct+2/$ykwc]0;]>G?^^Lsr։MC&WӡȈʳKLM\.,f-QUhrre3TzƟ(r/o:{AãSW/~'_}{ꮏnAɉƈP nQ"m ,2fY56DAx_ʻEFY. m1FMNr 9rM>*UiUr~ɉ#Ե%Z%oqԌ.=9V?BWlz05]Lު[x:ġwT}F8 n79q(Dz@F/MN<"ZiuR[哔)8 %HXfvtѥ4L@alS2cbD7f3ɸ 6Rrvhj(Sc{_n5afvcN*cIB^Lx >/pLk]Ī:-!),$)f01mbv #9V>䔑W}j.G@EF <ܟ3>ylCɩU dnVI%FIѧS?=:#P!B'5Il6) o sLIȫE= 8IpuO6sS~w?ۑ_ƮKëQAl߻OzQTzy&mfsX|..機c2!/ˆ?ɅstgTjVN`*׍mz6 +2]P&,5h-M4 xb Yؘ3-M2Q MhRf22J$46D&Tf:nҮ9|d9,PSC0q>7E/sxRfa`.g->|V˙}3 8'@jgyp5o?׼~.ܙ?/םE||uQcM^x,x O v\eFwcU3 ^zo5þ[gpv[J@;wae ][q=SXg}j>{oouce7w0;w,W(~T1?Qֺܔ_Ѡp+-fe$%ve2'6|5D*G3z[) |}tc_ TjA[AV O.'<ԘA=|Z6a[cf0O&Q4{){[mϣժVȿ]Z\DZEztnTT\C{qS)޹_s:nc33_P ̩ͳY 8.sy1y }?(6wK9ϪEzW~UEgUhw1O06~Ϗ0[N38I‡Uo-p_'Nk۵Ϩaأl8,a4prPg$I3J3R 1̷}u`y?zaEJJ@R:ͤLf?8ҘL%y sW4iEVy PIב}Q][-0#7Bn-n +1؟vlm|$p801U&2gD7eК:LD Y |(W*k'ӸE(.Cɓhnm`FoOQmPM rGXV'QDT8>?^"XC7i|zA^jd`RETLnL M-g"ɸh8f6vhʣe\/v5w˼7;o@4p~pm_`  '3Q5V |8:`Bds4@0 t %Ջ B3 A&.G?=v;[NQḏ\>'i#}‰GkUʉBAx%Id,H%A˒TCn8 70Xզm4x@|o< r̻O{MJ/lRMxp ֻZAHRN=Q^oKTJ1?TeQ̈|IxY6J%_H%BGq&hȤ 8&aFls1wɤ-Q$Rf cv j5%h$ ayS!qz S #x_o{UT%Մw*t&0^`p&7`gԐt1J҅E[muwk^,Yf ۊD~ͮ܂ܣKSb;>e/ދFOVyP?T"RC7N:NO04nWRȬUԮ[wAQDC4jV7,J ?Y/N<Z1w70?D5_察wjC uaYO6BoJ,"*ҙnH 7S{F /Ip~?0vp}IvcLxHV{ ggQ{%]~<#4tL#%?iGY)zWev4LG hQlv hp -&|l~O 잟*<(SHXN4qQH D>.<Ηlt\ޥwH oNYEŸ.!SR,Y^UPBZ2Aʝŏ2A X^sN@Pq.ACNd$) j4Nnq8ؘyaR9)'dؓ e:Aކ,jiAQU&X O 4@JMOUR`wjFj+buTS`?3X+๯5IźstD1!Is׆8=Ÿn_qjOE[fFW :pſG83,y(]H}E{kxgތ6_VNߖ!,9wiߌ|1GiЗ1U)qw4nF%5Uf͢Sq\'3Fyڗ'd Ǵe9$O1Q`نDnMq%U!ݟ:_`KK$b {[o";!<;"H+CrB\\UXyACNX%/)!N:Gr2{& 'hI٧{}6 k ό!CHsf *W|4\I3&4XSV4VTD2oPz[HZP3~ 'Y KܘOwk813T]Xӏqaz v4)6gPEA>Õmbph H~c)ڠ1E5E;JCP)~" "SV!<+f3x[nIد#!]OT1ȳ"c *l~8P_ tv4N& mk-2zgP͓{}fw8MM55_@-~}H<9h NRWU(RV! -rsb薇"x(4h *܅UD@rE.{p!haB o)K1ȋ&W69>-3}PeQNG8 j{ YL[?˻/#zvxY;9g -1 RQJJz!UQ*tx~(f=b/3qX˄GhZ9)4lS݉-뵢.:hۤ,IF^Fa`M7![Ogc'( ">,h*\ mtX@ˏPsFcW Ewͭ ^|A+; t!uFWj :mD-ItEUԁRoRp" A1hWj 9)ZՔT8+&hhb㲨LIVE)Z1EUAרN[3]ߠ)@g￵a(͹JOmdOEoҰKꏾ?4tO6@+|lgMXÚ 3ND|rx7!v4fNpq :"v .'zsѼ~8(^8D*Ρ] *kͷQsHUhK$) g<^T`V|qHR+$m ֔DILw(U鞯bV E?lLBNFf0e[tq=zoկhV$UmtsW׸Av}O:'3Ae0^InO͞l;'a\da,x.+R0$"Zn3!ˇp7m_cFPTj` ~u m 5U5DGmN\ FߦFH?f}L>dغ Ճg]{ t<B *vxA?m滫i$xͿ۩JIYA,Mf܃EٺCCw̸]A|xRfUZ-V“_sH^TX Ld.qI#ț0ђ)\Osa5?󹗘 4lOHNɥV ww{knҪBЋ~N8[7ש3.8WlUv\1^*p\/w_؍Ô9(pX)tke$ҠXƬ؛'ΙM%JJ^Gcb9.2͖nX0|Q^vB}]+7jgNo)>9QK_?hdހsR+6ts8f)8^Kݎo^ߟr7d~+Zhi\E һP#TS|J.!iJ{/"0gBmJ{Arֈ[~!&޷Λؤ_pL L@])=Y0yYURII>2KT=q(V!)h (IJoZl׊/#~`)E``B'~ꯌ??gǥ_I"~_$Z7QA%Ƕeh!W"(Rmz mjփدcmj4sT֌N+ʨ4uC 4(QT1]$ޗA\;]#l ~?JSV@g Aޟ5jP99@0-2䅮f>Ξ)! _Ґ嶖rS_!W&ڂzGJ''QJ("M: k)Zo*K{N̺0;qiJM ,iYBIYei0ͩL*oтnKu``kB j2*R߸IWUjc(1EY$ '<M->zb/H.})0kw>7 jgwJ R)^R8Qa)4aL)}c1+\#QXvҴauAat.mH,*8( ғY7GKaVc/e-tFTen6 55kB<)s9'Qg;'uNJ%9ǥf aIJ2I<(h ̟`6?.ռqjA"$ĩ,FZU**Y8489bp,nN\#1H*B??mex* RVykӤûnVE#}cmUl!KSW>,^Afm&z 43G:ŨXj1 *!I81wv5Rqل"綴:縕9K+",R+R(H  _#ZKiWfzf4h !ə26{x]Oh<2.lQ^Ù{ +S\HDU*XA\@ϭBQ-GV&V&`121A9#nh BQDƫ^vt9Sp3z\PO/  3/@EtiFR M ''M$Xء~Ϩ 7z0lGT;9WO组!bmBB~ uFk]\S yUPb%h*jK*j?/xn~F'h 6/e襷-I1'~U%>f9MOZ/DzG*r&eQQ8*%Sx:hEn˒c'q;*"qx^D3v+(7ʊKr1Y[˘8ø' TKԔ :B vvDžVG߁sDɓ1O |=llήobNntƌ#h-=i+j\m-(1(I/DXL|-A瑗^mYӧ0. O ET)9B2VI+j{X}s"v8'xB<ع/UOW FyY ථѣ)| Bu!7k?h 㻥/&qAXhr2U=fSaVp$Y/ O({|6($69bg6⨘) 0z 􄤗FZ02; 7Q.abA8{Xd7%:wsޟS3zDp!Tmbq=Զ*2nRg5j1RK'S×SJمU/k`CDC<=u*$WX oC.^!Ѡ1D>rTf^  U;1\iMO{j ?GzG&q[-~y_7\Vs,g?­bMoT^|{~G循.}?^4OiJ|ak{lΖ?GN~|]cwL؟o tQj.^jpE_LbGT"zA^{?, MXzw::ty 0EEMR&xUh8$GR@0rZ+*)=ٵY/tYd!HTa-+8R@hŴ5@>Iy[9b@HyoI08dF $e4nQJ\TŃϴz{HP"zotW}zNٔ({DžN:hM4Q+gA +%aӆ")5K?e/Vs!kvJчg"ysQ G`Q8r$jgxCA/ `|ֳyZ3>3%d rt3]"7\yF,vD908\AWP* <{T+7QHxsV+{ Ђfh^g0,t xe}UCG KVmQ ] /evf|MzB08*]}}v!6Fap2\#m \ h?S}I R!ᶻd6V,(C ԝalBbP|o%Z0 2<;1`[ 6,A 4*A%hF$ny0vY0%żo~QbFapȓ(rb!['ʋoyf ,T{(j&bϏ53~®pCf8 ^=&QNrw+E=]3^;lat7yS^zw$vgKQ'KBΐҺȻh̤s O~7-̀=".,iƧ@EssLOrlQ䡮Y6A9a/1sCvFɐGdRR%z3C{ 1DcOGJaoͩ:hԽv'C*aZZAHKF&a&N6>T#4*")P[fD:bR*Z7hz,]LB08:Y?4#kϰz^.π}p!qhEaicjh~3MmpVaهKE1Lt鮕/B084\LKɄ2Hi89S%K JcY( 3OyNU8("_ED)V( nPP! pq~`I08I$G Jm/7N.{/"&ͣfu2ǃLmHa9[h'GɅ/`>>$c8\aH~㙿`0S N1` ɰ !<>4 9`{O@x !x. ;>ƨhR5ܳx?oljNIrb_pI 7nҥ")x 7 M~j^[Iz_g =L7>m07w%x8xa82)1RwSC)hzgQ?Ta$|zA&CjA{vzHzW^, ?a<>͚!sNhՏ|vJKmW JBQD8-~2]o>K=F#cTNt-" 5M \ Y?uxMB08q$9;Lt:q謑CҊ[h̰CgR!I<Ř_}0("&El0( @ E2[Rr]RzS3 ;IHnYEUeSc [ luP Ut`sWEvA B087t,1h 5;-}6 L_% P [y&gr5 -4 !ěfvgIB08<˝2FgΏ/ p2bR%?0pʢ{]Ў&a] gJ%x5NjIpG[hGϫyq;20(]b*.֨]ZFap4X UBIFp u -4ʂpɯJwIHIB08B߸r2-+R t:m߻<+[>hq U{$F#5fZc\+GJtJ?t*kF@'%u ۦ.Ա}).eVYtg1}0$rcCNB08**d*y~/mSv3PQ'%ct ^}WZln^rl cFap2t_ڹwV*m%p qKaBJ^ ^}t;=72:CuWH*9=K2M͚o d0Lڴ̥cf>Q/ˏ~>q-*m5h%<ڻFap <_蹟|Fap2tk%h2p{3JjRa2٧lwJ>KMć^ ,sv8|ITBbH*oߟtkR_|$IK[ze5Ӂt}0va02Dw I uv9;߯|>o>)d l mP!zO$!gLlܝ^n$Cb%.fpбAfymi[za~hr~!0Ѐ{p޶1٣OY>$S%DlG0IxYC38CjM+&ޱe áNJe|Kz>}ђ*ZMcاHKf8H KcfZ{'0-E9`4pUNG_)-U"`OO?4s 0v0q q5 MWoZrT0?Lַ,aLa#ʅ!a{;[E?e~?ӣŜ>Z bhg;`o, oWWNh֎¤f| oah$|´^Oߌw|V?,~&κZ ~`"Zէդf~5ͪk)jd= Nԣ ؟W Շv"'@;G su'18;W3;nW `eP<*  _}uה|r@B;^F q6oeZLY5 J'sc$׽/`2nfp$;p\Lb٢%2B4TکpG@ucR:[+jvکoi,;,{$S~9@=DH84yACPR 70x5d`I&T1JMy8^G;t(Mn+R/f5?kDo` *Za嵑ؽsw:YcͺtH0B]!@3Nc \z45OjɆ05ΓPBIǡրZaŠcQcQ -8˵P/cQS*ZR@[q|'_~t)K nol):=Za(a(a(ah u|oVf?|[P}L"p ci1U[$\),u9@12ļv\b~P4\k_.>3~=҉}(awߞ&N|p9^B|R@T L/~բ~‚EZ)xZGU:qwLaVy!^[T1WTj1´_E PW oiuOZ߂v|3ZNi]j̖٥&IJ^ӾUuZ=-ZazԿWXYW߼m&qV~:4'.[U>El^j@#tn+.bq8tz)XYҵ_LZi+荦 )3o+AjPCEa: rZ.P_Y\ m2l֬klР$?:dv֮m8v$W1bjl?D15\-46L Z-X =la4E󾍌%( ^cp$^PWsTh}fg,ۚ-TXDt/P,IS?db~U@AO-(%w*V? a!_ } NOVoM]Udէÿj%Y=<>m2D{ԮRsm9LI*n=~&lo.|p,bƼX+t_NDy=. r'.@JK c r+|9q??7R~IՕ|9LxsWWqJuY/>Kuc\JO~ȌY4L&taTǭ] &(ի=֥O*:UUq15n,HՕQS}3 6Tw+?Mibb?h!Bićm篎Nd8p'K'I'Kڇ0/OT0߻=Be,ޣ;O!pw1{, faPLZ D0žō4)#`^3[ᥴe|Y78{IHCc2w!Hs"{>eIdf)vs xՌȊج2>ZO32n^Jk'y%syJacX^);Jˠm2B W^%#3+kgxCjǞO! .`iՎd#HX>|9g6oSZ`>asm3oNBBd&y ە8 ;O!b_;(` ƀ gj=n#GB4qJyYX>M"eR}#HdAuz<*FEdEDFd\(g4^B|%Wj- 񜁟1CSإ`01/!iyB 5Y0? Z[j~Ֆǰ஋WO{S5s1\<緆<^yM__.z玫߬7@sޯ[iVO;,s;חq׊sVHX\gfZ SΆRPgvQc. FT~$Z5!s4ÞS v$ց}Ұ xBiAW+tV +|a~amSogףp}H炜,pzyi|adeS"{En _`ϣGןpkrsl9_X\/Fg^y >X9i/v\X޹5iIʃV@ur1ZdhNOm6y%7g{;Uq4/lr\1./WihP &5_M-PKН_z;ԇޏǿ|N>Y a_D@gi #}pX6yY7O3 i؛MklG@Mbbs!$X(3wـVJ +iP-O%@{~]B '3NX}~o6?>y/]JOʼn H`6 9a+(2\']&B Av{"AXiB\K'8Ǥ9G4n9҈ 8;@(AR*FL" {0Z ̻@(Atd% +,d~IE! + 1QJ \siyc!&PbtPxMTi [ %` 9rA["d2d "+]:4(1 m@ljAE@(A1|L!`n4"!e lqɜSjti! ,ǚ֊2r-"ыHH24xu] HYZ`dA\(8w --A`?}o_zћ|:׵8!~;)0q#qpq9/Q8F R5 go*9Hȳte ҧ^NA~U=Tx>jF\gKY†Ѭtֲ5&X|U oS<A a}- 6`J~, 3[BOs_bf_˫)n![E_L301ƹt8ZIz~[Z, %>bܦkj J1*y^Rv$w8&ˉ~~K'\r/PoKǔej+o߼Ϣ{u"ڋ=ȏs 8.*aâ8bףG:2DQ}obhmu[gUz I3JJe`skHd"tJ`%&47HъPm279wWm~űٰ>?O|m\*AŲvvj=gfkydp>6Ev'G}.(a7V39( >qSa]g5JPr VPƄp&q"Os):TIY'~Zqȍ%Mo|.({L?psL-zd):tGn5X[CSO7X2U͞MTWWB<~ݯn.nn~bٱWuuWOԜks4 5[#cрu JMw6383l\ Vne%F7Bt{E&W휩)0vAw Ȉ<==Gj Oai;|#rg|T_n7e>u҇@7L1`<)| ̶96 t6U+vi`&\V$)ҁ#dfQ3M9O"g^') \GkAEdʳ &ȑh ĩ|2{4 ͱNEa:嬑@FǨp!dl,DĐ1DARhga,w 6p˧_(z^6z.C 4 K B`)XLE X TDͫUaGN\b PyB(wm:`4}wӕ\BPrᔨdLJL2]%ӆH(8qvq~ؙ-Axb#ol߮/SΛd[ij1 {Cb"\'+w p ƐVi+w"]L"ʝ4EOc ,,|r!|COCN{*x'#֌DpK)$%"(PZ0֦G9Xcȏ?=(!wUX:ԗxz1SVnA-hqU%{;] ^+UcfUw)\1$r2!9ⷬ9M!昒adft =Xg駞d5̊5tW/.i&uIUm1U?}(j[ͮܪ-xqv=],Fo|FMlgߵE{dgzdu}וixYM:axY^ցդ:ȤL͏)fXMS%]tPq?B<{y왧8Mg:SgCnT-"'֩iujZoѿaZHRUGNMSԴ:5NMSԴh >fV:5NMSԴ:5NMSӞ0^pi< -O9ɕI1{`ʁrʁ6i/*GW `RlWENprdmI~<^[WD;+4^_`f0VZXxhrՕI2y n?>;UEj0RQHҾkXc.]z$†$Kqg52r8jR_<q yJ:t`Is5wAو$Vn02DŔ^;waՌ(zøf_9_f q5׋ܜ9d2dgUͪH ?|;o0eUX)a-@&ztܥ֓7ӽi)k\ۙi}c;q_@|ZV-Wٚ{뻟AS~/L`/;mD7/lv  x@wmsv c>Z\kks}ϵ>zv`g7(Fd}N}ϵ>Z\k0W>W>Z\kks}F ZCyOy sRN@sθ\*EaO>Ncdj?_)sqz  nH1+X춼K3v7a)ygQ-/$HKK܂ňmX^κHs8KebV"'u/DRHrY$4fnY,☦&S+fbA=%:ѩ2Z A]yDǹ3%2RH#M.h7|<dzE͂ffK7`~|GC:L#Sl8ݿ9Vj04E>8h4vwä*wW㤸nJ`E.,4\sP<%S'rJ,ƹ &Xq*qymregήP ;EŹ"H)DS/ƌK 3&1׌3e Ev0 8-iri196J%18}fV(ח>-U1Kl]VԺ,9'stNs(i4qxgq&lLI%G+tb"vPYX6fN\ᜋI34v#xFtFW1W+߅uRX'w{ZvIL]-SO&/βh8w F9Tj˄pZsY5-]*/)\jSkDCf 8AvYp%,NcE#,v퇯,^<6FI< )o.W Zɻ(= PR.=+̈?tp|,]ic}z׼XlBY<ʒ4au>6n{Z`Ϊ }Z `65t>QRЕ(zuI}`6]ՂhIm+ʮE3t]*z#BxCW[+Dv-k؁BW`H k ]!\#}+@+H+Dd>ҕVņE!Q໿p7K\m]+!W::D*`zM{*"Z*"JÃCWJ˥Otgep7thE瓁 J.E/bF`K>7ZF\=c]@W*E+L.7Ut(5tCbRj<+7tp--km=]f,t5D׌H+/th:]: ] -i1UAd*ot޻`J¥zDve)}kЕ2BX]`+k/thS]+D)T Xm޿5^g?qW "Z#νU{)zatJ)lԆR*2Kp!S q뚌Bd*s1qySrR\Nyn\LmlBtQ)+[2ZldX?JY%**qMX8`t Py(t狂`Fd<gvnnq`}sL ?q @/]˯(}s~5PKgRbeF2Ֆ|aDe}}1 so7˽9EEvyC!F gk#]+k/th-|H-!Е(zsIѲ:-m{WR[ --KB)]te]*zLfGt`BtdzIW05Q ԶKQZ]9ky3s"\ h%']9da)'Gk Z ] b};JWR#]I9i ]\AFELv}+E>ѕ0 ]!\|+D;OW<@W+5#Ҹϩ{W@W}+cX\ܗ6cÅԒlX3HmSa\3둫Mm޸$VPZ*CW2C E5t(YR:lEK‰'X\r":h)ul =x* wŚ+,(.-7]+D)i]`{Ͷ Qm[N*Usk򈮸 j\ ]Z;OWRuFQЕ\WzDv >ҕJ0kd\}+DV~=6xW}+s]`s Mm;Ko=RwtGҒR.f JuB#]#tq+[F= oTWS#1>-9ƒUfYJdU{8Z ^O6X!TStpz(BDžū+FK8kEMzp9m&ZN(UNWg]"z#G3ׄr4sMw tGb yw-a¥B,Uk&V5ЮHWBǏtp/~<*N~%C#]0O/?03{\)/i&VdIxN#&1Ҙ8C>~p%^V [ p|Ά|Nx6ۨrVٛ4a-.Oi-#+,,Ns'~]7xL%2 č(, b/bO8~8x;Gw,w9c:t~8GAQ˟ޏVJj[Yk))uvTp~3l_Grn>xA6~F10H}[7qJ+M䦥T>bX /pϚ$(_ ѶsJݱ$n;/&ѥ[}1FV_}تi)Q4c׻}n_.$vcmUG\rpN g]ﻻiuU-vc1%Uyс{!µ9F@4cN|jxrU%zTn<*dq/ɧ^5 jʪtaY@}&gI{ɭ+ LټA0Ӌ6͇@ӭ,Ue=K:YK*_2&w;3C%3:5P\y8I* Yu_p㛻gf公/$~u;7j:;/UFvb6+$Ml;Q)F[=}v]!7C˚"G }H2ZքL{=MH·෮OsV,@Yw~4Y& Jjd̰_e4(Yw^ Yʬ@Erup u{ҨĻK!xBJߧ)F[.MiumdJͻOigr,ѕĔRr$rE삤[Ԅ䵊2Zt44ta,+@aWoMwb&T{,fcsۤQ¶9IMV٪2B)\,p4h@4cxľ =&0 <uSrpLJȽz 5t8e0\>= :NՃCz>Օ0ۢci/-^}&%@t۫mՓ@&Cgr &7 "ލu,A|WQ,php *j(;tԕX+ p ^|? 7"u-YoD}J=equU\|5egGSJ#y PWЫ' K]ŀBG]+.^ԕC?C2,֋T $zutXFTQpQZ~jҸ$יGEshEODkhs E$ZW TAk1+ _ob^u gwl6)2z7UEi:$l dtĥB 2.QY /dzn@JfC|!Ɩ* AE_JpXS)t0.B-0@,E‹ " Ee@-UR "PK+{4YS)2.aX!&_xb^:k.Ct.<2Ҹk#{.z\OHmiWh= 5QW2݁F#`gu#Yh+~nI $cuu(k`(CԘ@4)OQf`-SGmnU>{ŀcٕ6/_0$}&>M3]49!ML΁ ¥iBX%:> ^Dve}*mOs> fWȪ^z V#hW Qu>okBHF8OXisl9bc[Q& @7w^AgUeFYPwpzEн/Z-B\&,5GL4OLzE # bE!"J$ҐnpIhpgsn"eg-j/BǗ!_vV2_RZ"+TҎy{}O2V2(Ӑ(j@I*8MPb!43 bý,N󎁞)ZԹ %v0@)H(8U2V7C%!dÃlq2XeZripO-˓Co1nICB^c'P! ;6 D+WZz!=Cg̏ڦBD0IR %, 34cX.;Y3X 7X~{p@8p?Xpj]z,J\z4|ˮ CF\{\i^S nߦ6u`AXebi_ؕ%eZqGr/օܵi[53U*i7=wY:u=&Cpc.bqZ7QA.#uZ ݀XwƏ>/1&E)foYyVV? Խ$KǏi_5\oVz!PGJ4oӻoQUݠƩay+)hl M3VCp(EpIz OP<EWf05ѣqSB}L8V+oPqm Qo7^- BɎeMz'y=~#L+I %a21O"|= l/ghЯO4ڞ fTFv 88֔uwL(^WaQ F{C}6]z sņQWcrk\ Wi?EzB\|ߔ()X{=u罗̕!4fus={>`wRp;ة4jq;%.b~I5P`Dg]LQE@8"&sΞ` wL1o xsD8V_0I%p:5YA+V9: 9"\e?!1Sp !?7)\g~In\)%+=ݏJ8SPtӨO!'X@S) [oMnJGZ$RkM{ðZ=1a_wi,]J}gw *?>5gt*Iq^;UJ}rq]!֒pe)[?@L`W2L2S骷/:Bv]?pYЧ=>CǕv%eF[ě=&|pCW9tP9cwNw'Y;ʸ U!}~muO;]7 ?$nGL5xbo=QSZe88er@@NX$<`uƷIg."^hJxaQV>rA]0vFצѷg|}NY8O2G/Y(Xڷ)~a o%YHviˌo4S) C-žr89V{\iSrr! lUؾk\ K;+~R78qa8ExH6W48%bޫU&YomGߦ0mN=}ɹss'] ;~ ®TP}"E+Y:e:<ݯҸ; sAKFc}#4Kңxi?Arx>>V3/-W=剋=>Qֵ1Bv^.xs|r>:g)$w?5095}_VW~k&ڹRB2X')(|E4Mُ$q>>NOPΖ s0fHLtQ#H]_[CkC4J!d?d H-YsT&_)~yMm1'c?.b$;E>H+ǑV#Gr\-,Ƴ$1arL1L3Q) H}d4q"yfic!%Q/͗"Q(߇bꔗ:F/e@uxT. j7u TV{y\~ i͛ܧcϏ??H#ӟ7O Q`B\<<,I1YF)YG24yJS @,D֧FNG͛k;LW:cRRO>rWҒWJҪU򚎠bЗil_nQ/>vZ>Oir"S!*e-gHIbaփu"}~ABA9Dk/OPVLA=X 7w7oOI~=v0P}zwQzT:PdO|7?ŷ|>7+qLg:tQx[(3VVe;GUG"7>'}6X)^AKqi =$PS;C$ӱ&mzw-*u6*4N;_1He>< Lx\*(PPBȄ9@Y ! @hR`q[jAYnCWP0}Edd'&.NyEJfs@E~&t4d!# 3%0̈́d)4W(JRŐqEB<HO E)m7jv 1eun~P#hf!# (C a绅+=+Ĝ:MTu s)^B)eߊvzE}5bͣW{=S8lmV'H\,8VQ|tsN6/iVubU ^ŏ?j\ld!k Q:xmYm й/H_Gj:{G)̱1(I GH ?G(WNjj~|n,M9q|wb kn%EtE]~]ll rD!zEM0xY RO-vp70D+&f!N_SBM_z}ӿ-_mdo5Ϸ׃F}V |ן/"?䭿I*,NJ; mj(FH)yɊ< y m0oMZz[rf@v( ‰)ʒRym R@N{CmKIUؔ eYb%j,9ZϊXd#bRjdg87y>o,6EΎ.3v5rr~rc^X_2UvBEVY%)ok 銏A1RV Qkfw,@P*EoO"Ԍd^$+,L4E0 :a[QAuBM=Ts֎i !|l CІ>2XG f2}BͿi6$_¢*<u(r h?h吓& {TPEKƃMH׊QI%29AK4rlބ%HpAWx8xsd2#oM/ r8]ۻ>7,3:7FۺZ[,z Cg?e^R¼f]vW{uu}sˡo?x3OH??ߺeCsq w{u+kf#n+?O8>L# 5IAuȐ!5,HlrM(P\ 1d%r8a4doH(@*Vl0}zɉp ^p/Fbq,>>ex5!ݱ?fO$js'+6{w8>Ӑ-Bj _nϹ8v~mn)k 5BYȆ,MJ(eɰAtoEM|}BNc`˟;fq?ݱ?zP&mUGt1|7Z35!<4tNT{hێF-:"+=6VFyhg:g8U:/o; s~TlQFYk"ɫ!"x .9TJ7rVabJcd9y3RUe@אԃ-eGD (*Ig3Z:7JGcW(RĚd90V-D+X 98sHa 9HA(f ?NdMևH %` 7Y 4\vֶ,ϘLL )PjA JNkSo5bǮҥ|늋)f޺W@9A qJJFh =^I^|萫c_>EG꺺ag2٫+^=߯R59v> VV;۳/TŚòˎ[19DLRX š: Tt]^Sl=h JQ2 +ySXgc )i߁} P HJdSBn~S=JBfC̚pؒ`jK'SVX099kc6Afo B'@ -fX3P29lk*6tXw[ȾB{! ;,,S*((2F`dDTEYQ!vjM|96Io0m*mI}(KZjYq VpD*vWCa7 @d;j-.؈G2lͻG6Qpdv[o{m4ԛ_Kd=)ṻcǡls#fހƌ{WɖP eea8n[۲0$mYP^QtlR ɘ[ԡeaODe+;]N_\P" ? 9о5vʀ%C/YBC4h^;+ZL~8 p, 7W.+e^Z1f_Kť\ͤ+r(YU=WY_gYD֭Ic$N-t:s3pvbu˴ju׀;H.F/aԡf>8"ف HR=1 R[ᔶ>8٬)zڐ2,W&&E@`}j ( YaE F, \ "L^fZe!mm եE=PqWbȀ8~RZb lqs_LPGyґ E6J9VBF_x-V~uQ&O*Fav[2ֈ%`E!󴋙QL#qZGȝI$S%rD(kLpCSr殀,6 0M!{,u̘P\`H 1}",Rov;U'mY8c:3紬jpZȵq̘zJ:~~MNH,'36.J7$}zI,t,B3Y*3PL1]gb ΄h,CٰH8~:$TB {&xCa!p`L6:Jpu Ŷ8Nj'n|$r:<',۸?5HVRl?5brIe&cڤ&߱'ժ4]|iQ<^Lהdq`b #)؝i%i37[qvFC~}'룸LtشXnq\בxƞ{vXnvc"I㝊v*y\gvn5T~X O"+۷~3'v~?#ZT<f)Ӡ8T5s1l źIl0Ll7vs`͐$M>0q siږJ0XpCY[z%m,z8Cܙ2($R{RRL>FaRP* ]!*R[t.Qх[ئQ1 k ΢ꟋXNsJiJV/?ּ͢*o?eruqv[٫aEMa/@:D~YSҹ0M_':xuD 1q|xOmk`mC5*mIf"\/ bw <h3{?~?݆ h\KbR ySGP@O~ZY&w9:G6cRt VL c7 Q4:kp~vz~jz` ) hfJ0g p7|n-VotOsKE;?+W7 g4g(hٙ8'\U?!cӲ^c̦D#_+s $8IDtnS|j1WN25o6 B %0(0bS}38;5y }k7Pӂ}o4ɞTPwdlgp Ն9x))dDӸhOȉR Qݥ}ӪzuuJ\vb JDHY@~zl3O6-Tz9؈-M,]I94w pSJ֬ bO񪉈2Z65LP||?n`f"Ot/7mП:asQ-h6в̞Į)>b 4%V:^=Rvkm\Ge7_3;g '$ݿ~KʶdSH;r8IK*嫢Nޗqg$un(Z8Qi2l@;po RzdKMQI }Z]Z,G .v|9Ү_ )eiB~-q_5Е!ZœCOCb)Sw{4o8hz(a43O5j1uQaOgPbc`yzhO[1/6)йBh)r˵_*/2iS :8lc@)1OZ*+_m.bG |zkU Zr^M=P*ØL8ѹexցp~;jBiǝ8rEc$S?Y, b5V.ZO2̖Fk e}˼X0G#^#5C !~ӤN}1 15d֑y˚2Q8OᏘm߼ \by@\1gR0*aRb<q5BUbt\2Kǽ6^!v)a׺c$'䵏S |LV5/^Wˏ]45W.fS7#tLPRTpн%sZoK$Q ^ vs!uem 9(-1"%% 1&oؑ][;^~Ka(}IkN HskDFiZ(jquxE-VR' {3A6-sI>]2{i(jV907 I',>TWbL xzK )OtV^/8 ҪQTxq^A0Ɖ.|U[Y/!M`_^M!ٳzR&\XJZ(\!Q.,S+Xz| nkzJ?s'3HHӳ7nXԯ[}ډ#tr8Ɨ_ 53G\X)^MDž՛6j׈ T ھ5T8e}dve|I:G(20 =D1z db{'5RrAfwLR[@ 22ג!9%~Sp$EƸ -4ꥃsֈoM{زDcq>,Ǡa4F6ZGEI +/WWjsSu6QƐ w^[0SD%T~3jŦH٣5N#\E;v^nz2>}ܠrzMaAbs.60ȏn.j㳈ڼ ܸDu;h]4YG*u nK/wŒ"=}0|ȣ N혬Q,#1ȀILMQ<ӟŭ:L0$8é?2DDζ!?d=K҉?Q̘TxW0].lec_߁ۼ6D& 6*OP{:邊=1/-˙*5|?& JD49ũV+t(&LF9m2‰\qr%Gl#%e)֢jQoXcC Hik)E/jtT (^}-16 Z(K@?k"80Cqu).R{{˨pqrmGSk)G(#!(&Þ4tPLO>%+s2~\cXӼ&5TWqNnc 5ntr6m]uUA@NE[PH'Ytǁ{HU" `HX / z{!w@p ɓU#1^ up[vA'?/>y* ѐ" u-%0p UʻXK4{b _zY8V):Lj,bM!BփGea!6IJL4G4>dۍhN *ĐZ^F'2s=+g=,%dq(s3HGS1>kU&)w= KD1cAWW07 "C)[-hiܥd[QF88s9f^.pNKM)O2*P|$ПhQqIR] ~Ip)W{-J2rqO'dD jҁOC=,1Di$oR-vS)F?/0Ue2x֊hHyt ɡ+c0ǔ7蓛ʤm`oW:FFyF|r\8H(1|DpuҰyڠ[Fz0-Ji˽mH'PNO3T$3%&U J+*/c9.hw*tDgztm{;^Y˜ )bS[j'=/6>9GyaMyU|5qiTiWN#& D(4Y8C!NV:Lfգgwи"7QS1a@3iY7!:hP(HOpFLfpa{Ͱ9a/?Eh |Q9a PHeEsCAc=g+iqZDx\19\F_isH૕H8 / !'5yM,8zK(8_Oe"n9DY?2Pɔ3ڃ2KFEV0P-;RҺ h(iCD=F5N6{F% \1KxE$ҊYdDl56uuvQiN˝.Ɩvs1ňClB?ޭ蟨$[o(F|]VGJA/Ԃ!SYuHS JBXXn=;E* m\TlhЃ (oޟ` &We}w}4 T1<>p\{TIH2;(H5G A%^v\8"k8A3/:!Sٞ߼J\O'C_ 䳿8t2-]t-E͘&z 3^̓SF+; ^ZWӳY4 3qڮl]0UR2'7Nw! 8qiP2^: \%zspX⡬lvo,6aG 6\}x@Gm"-!.7tFvjNW$@L20˨j[݁ B,<. $B% ؗߨJ_WPcXLqdx KPbrԩ/z!W[T\k@5ZE (),.mq!]A#9Ȉzuu>+)HTp8]B.0314qS1ap 2CS?A~pyaViSֶ0$ VBC^NRmt[ͦ!ɤAD{@A¼ !'$@lZ.'[T")Bl(CN .ZL:pT"eK ;hΝN!2Nbv}I:N9dž}ĕ3>L.?(2O~Ii7!/(5bBc:4V)GFf՟[Sh$K#3tH,nZnjC[> 4~X]kC$L?{WH0׆wh`` ?y0kyT:[U]:J#%H)vU `D/Z=0GXzmqqMul3C}>TxZnη($H14 <a/>mFl86 DrͲxL=)DOFZA9;oc "7}upaSI*xJrѽ%jބiR'ʲX C;U/6D8H4?j =H(ӚLfPVrSHMv[).(4bs=v>+We͡F6aG0^$Pğ8}cū귾J8JFTO }wmMé>G 7=2g ')sin4S%/^E}z` KlM23uqn<A0 !Dޅ|`ޘ/AQ HBomVnQkx>7Vzs֛1K:#O*.F8C )c@ȓ16̱TR* 8kDpyޓkxVHQq+4Zvms>;ϝAV +Olb!Ļ \ Y0!&qTLBo/D"?Qu=e<qڋ.t8sd k&h>F3E;~ w2c|y;Y۬0bpPe1Α&Wcm N^6Xc / bh+9ĥէ(b)zq_fx|^!%$KWХ>;K 4i&Ӳj-AyWAriJQ}׿^f]=j }T+Zp4/XY-UABK<&T+p*LIJ₎Ba^du^Ax:N {e| ADY6ПzcM HL3zq`Asm.8q',6Jd"+Wj<r'ʡo=}@NG <DNS.h3dT c- ]xl dl`}bAV_'/v6Y-c#3҈*Ue c1ÊC +/b9 #B! I)}M`.\gcm m{$M9>.>)&+XNY[N0,\d9$JCX) ~yZ:.sHKD%u^QNN'A`yIz1VZ^\"AO2:|H?@ yJ" Ci Ř+'+X>=).&2*16d?V#R!6kR:;V8 5ؑ^OTdpw|;Z4--C1G^>Q)X"UuWPm)՚e4U7~V$Oq)+Z:X8J! { V/0$c$ꖂu!vXX{\ba0G:]思 Q-   d&ӦMyrmeqY)H[`ڴq0h9OTLvGru v5т>ʄ"_+pgCo`GdN cTyAk'Peކ- %HgXq,HXQ|5n-EO[+jmtg4d} tOaɵf$yLѝdW֍sQRx2{0p#+`ԆhDWIt$9Hi.?a1Jx8RT{$b8AqT9>=&j:sTLjr8A/^;z\ë׾Tw5 n?P4d xg9[-b9zR6GzgU_8*m}<—JTjgߍ%-q'D#8c[d&+Me3gOBU,;/4ݍR YWOÏ3L4/Fi]g~-sI="e9w=wKDmk[1IK{sނɤȴaG,8H ٵ6=pf%6|g-^>D\IH0G3zlM2+89|n0r#%4FF=qbg5BO7>ԋ^% 0u,NJ߸{2oݷWAZeՠ8A( O+pJ&H׽bLsƎu`>+".[7EZ,\0YkxH:׽I-5l޴ϫ%/EDt.p\YՁ#땤8R0WW[L !!%)5L|9ƢxC1Ƅ^^A_3yR II,Es}0 =<- oBotF8Zʸl:4^ ?/*mn;Hw6VyO! Hj' K`x6nt 1I.c:* yq;5'lCti%.^|s- *|{dh$()$2ؖ-**٭D|R(/ʶ#tq cX_XOrfL)@u(҃;rRs ʩ4ze6ʲz ȷ3-J)' r:GpmfI,I7aBۇB^tGkWQ>6x蔝|+0a@]G2Bt2ws1 BSn`/;o9T/S^.1Hō/|o_&\) ;T,w{6zˢb}i^Ij=_dTfi<_-Һ$ Lu~UPׯh0YvAZ1HGp6/QR(g`|g^65q^]2G2?/ ;CX:0[_YI!FQ5T{ceXڬKzԐbB O"!K8C\hU\DԑuWHBZ+|͕z{\U?LGCIHE \߃$"Q\p(n".K Yh@A_ h+bMq,S ce[lj l+. Ts=HǑmab^eh}j5夸~JE.vg853lNI$X&>;01C*F0h#)$᱾6u'wx4 ~g˝@~Mje~ ;mWĤdbf͆yYNe0r4_hFZ/o\ScRo%uԃ|唒OLcx2gb|M?\~feFe_FL C$PUnV -DJ(M!]v `  5V-`tg cLJ%Xg"%*A)֋d$c9j9PZe'Y?A9_[4oU>g:6浌,tSo+|b%5lZ/>F3ⵍCȫYLzM~>Jm{sr:*+6b4`9Կ##$_"H wmmI~`_ !{ vd })QLҖ }):ϡ$I"lG7e 0~zn2l +RoƆ@kB0֞^J0v)r(~BCE~R!=VjQ"۸`7T!?XEue@+W0!i |9d{[WRwH?|EZ \|,C@mʰVQ zH]TBڬ< *Kg`p9#9-5.$XGn)L'cT'9-eB.m*'x[PNzka"]qC^=~M.8$ G$x5ڈpeb`l]pt.]Ϧӵ;ˎЂZbL;&ht_NB/ $'HU`J"W`,IGوIQ׷Yoz W(MmU֕u5`gųhC@C9y"}0 0ktf>rnAĔ"59\N7J`NϬ6,d2ʀ\^TY.;ۋނ`H&;IGM)']YfFhqd(D"C>FT)7'OqvZ}/AߝaupmH(ϻS(YG7i,P s#CLBZ$wR)n `y[ R#Zl nu xa !bƛ+{+s*%uׇZ6t~,&}֨zc C-0Y8Tee2Iz Z^+@9|8 1uMm5p]z"Iɔ3`j BF I[y$V܎ ]n]5k'ΞkAʄڢr0-؏.lbRO0c `tj$3 J&Qxj4Y%>_OQ du vH`^;7TgBLZ,NxW݅r+t4noB_C XBNrcxϗ~9;.ƠU.5h7 #(ϝ!-C8ᨭUn?,0çpX]WgWr~h ZI;Z쌑r81E vHsZQIk0ZwTLyT\Ar䪝W ` k%z3i>Nj8ŕJNOF@f & g",#۫}(m)70c|,)13XI=6" :8Z0K, ϴX16޾𽕴zwsJᲇ ~9!^ dD IXd1O;|2) _޳z4B_qF9FB0$Nv k@Q%sDZ]oR(J#AkqwxP`j:?TR>1:"ͽ?DPR d;`~$`Z2Av?]WcFJ wnY v FLIu_oݫ,h~LpP+^K)}o1Y8A8yn28(c n&싫v:t&JW w}gM')\mRU݉q!k0^ hG6\t8ag#IT@pٙuKY`℗=Dќ{#Bnʯ1K%8gkY-TXOƷ 1Jߴ)n2$T=mD2Lk0v =mW%@{zD!Dj}_tǜԤ@Asy׷]E{j0^Bԃ;N$ZP)::%a%˄\,0 d2H/sd6[ `,C v9r  uޭ>C_6I3hWZO~V>leޓ o )0,/^OG Ed:nk?@8y.z⠋dxV[q ~'FaSE-[dh$pi sl1:azK^{8a{-Л}8'0s4osk3> 5x PRjܱL89<W?A,I%QЯW-Q M'eMR)w?X+{"i6K1 >/`r GtL<~{I\\h|("ioe&y?ϙ9 ^QKDOrR i0Eﰈb*)戔a ,o@n:ˆ.qIdJL^5p&DFԘy-KFp6>6)1YБ[kHلR7vG!>+9W}$II>eZ{:~6RQA4} 8]ӹqt ;hUc2ϑY܎>] -Sf!UDLuhU燪 GdZg.4İބ@_,EtV,x-O4[/@2SI c,ZsByNh= Մ֪gKK풀"6¨Hqh)w<R0E:+سg-]=<o.SymhZWH.hm%(cNLQKlPXFh$\u:kkM_[SyÉcsoRͬ9Cu%wZT1iԞ\l(?q2L_~\ f|}zL,hvW42M h,eģdGYV:8*Eh+ANU5g`J{+0Wm8@ l<бkuX bAy)†hHT$A^zFI]hvӸ?M'4 H-qr1IHDOcΠMEWrKnʅMYx%eQYnrB BLr)2Ssވ[V"$+?Nm(FnE6VFR5ILpqM#:aXxmo (;( ϗ˵tx%>){QАH7QJ?k{Bt0JT6!EڜLY -YױJ NjgΚϦ9<ىmPolzS.t"%vcrIdkdD&>ũ,$kb6 Vgzڎ;2L9;NU"[ ƋtqJ{&96* oL/Nǫ.77WGw֬=5mrV|dv>ǣcJh%Hg@^eL IsN;ZsG$k5=TJ:,NZe84E0 >6kGYeyM@jz?rO٩ܷo+ 7l*/-[~6PG`vl'%s4B/܃=~M㢅}5N\բ+jӫ ۊrzoˡ?}6aVs >ǟìyn(E%h'ooߘb_O}/hkS?*E͵n߿JLIՑɴɴh+m#IE[V^!{ #O[ۃI]EEU%@wK$UŊ/2c\Z~i HZZ2t(8VMN|n+L y*j %ӈL T{Ztof4Y=: BEnZ\B/Yj>c7V䔕SO?nO){pԔVMeJgBdr+y{^g.#e uNp \+n|g{M}ݵpE\x '~;o2{XVtfzQ7cǝ?>>\3 e%Kv/:>0i1tN'ױW![.:8?~O˶F[= lS [in>=rBl=8A#(ڒ&M^>@߫LלNbJPzBf5JN]jġC&iKx@S%6)tnKnj$Y-5tTh_O-&3׷Fmm<]bc^Djd'\s7deS5\G؁آ5 [4pi4 |FN< cCFM,ΎexT"a[5h D@ IkmG9%FƦxSů ZvHV\J3gV<->X88Ij õF/Vq">0atM(zUR;+ _D;*^0ʘZK/ikkzx[{qv9;~oިE 7٪aFϤf٤ڎmkIr`L["8_ 4m/5(ڢ)۟:1 ҩ/‡!jҰ9l^enjb"¹N+9hm2>B3C0b*mdO4M۾JnڿvHQjk #7u {g椷tv0fXÚ*8z[-z<̀mUoPj6P+0!hK$6'oM4uN)Mi@bB-Fi_̀ݞZ=ل6晐H9;Ptl^3 46K믿b0˄>Y1XRJSl,)"zM~/~:| CI6}>$kj[J|i<_Ly|tE-=ދVL77W|3Pس1\(Pwnn~%A2c+K|f:Y~ ћ ZP~-g MH!U:Hdڮy N-mƼn.[xIKQ`u%g?ty>{ QIfv|]}8ΏB;;|zgl*|T,O3Gg]2m%/cKhJ!;Jo3^S1Kۆ/GNj3?M׃U_,2{NcU8*n%: #?rfWFk}@,OË*?.PXۗc3{3O k#HNO"ء}\ [Rq?>.6Rk+c,z\iAnZ-9N+2rjân l7ZZ_Ie b>03MMamVjj͠vhi߭]y&`[}w1UXY|.8B}x5>nd ]۸ƗpX'[;XiBT D?VZu*ZĪE}FTS?XGhPحd_|q-m_,[i@,7Z>|O|-T{_&ᬕ3cQT)/tXhwy:WgQ=NT@jV_%ЁeAV3 xgCo4l͟XGM|Y@¯檽*N.(PԂ*Tskھ`n-8ӥvNWqAOqv|);/8 wZ /D%b`6̿|՜-`pcG0Fc%uoڶն-tm _MUܘX^iq~'\~gh?OF50wl)#%o5uXawșl~r2"8kP3:$& k8%;O]F>2a씗fl$2ۜڃ E#!FIdKMS HтI Z!.-&|8>F@a;)U6%ۨS,ﴄ. ;S9b/yM"6}Eceh {o3$u@,)VQQy? sicX#3#abT- *?DC=O{]4f(QynKIz0^+pq6? L N|RxD];f,KQ҇A)!NF(dLj.m+].)`}Pky&_ /t|5ĎvIlU0+>Fi0LuY&x;UTs,Z)/*$1"9J,7X"EaY& `i*)֔DC2[i3PpFN '/&g=l- aM"kg4v WUPث.! ]G YAM tz´EP;L!t9O! P$B*Lz #ɲ286cߺ`1ot'=%hY\gcb*Нw rtҰ.Rbb zQKl/d/,$I䙾69SձPADm FYW~*i5?v{N"6!N9!dAFd!mDX%{Dr>p^z=ZV¨_@QMsRk\OxQn?tm.4|!8{r'n\./rvy&އ@7J!s8}V_GxoV=_4զr@\Ĭ[yKF)j_w=o?amDy%؝|%VR7 ݲy*^?O$ҶѲ 7)m5ׯˊc陷5t]?hOCQ$[0݄`JB VBPXPYsN- כhUR޵5c:;SfjYXwZR?ZzF>D`+IlawMƶRb=![+8j|2 DkE#(zUSl gGY=_cN.C^)h5Yj)SҐ&`v[3Ήb⿁LafQ.dDh8C tVuEI]fX2dTJ6߽ R@40j>ΦR#GVnA1"O"Rn2)AbњLtglPH[򇐽2/HvCYo2#t0MpQLi =̷9~í!5oJ:h9<nj ?R_bcW2cn9Ed3v|YRz}|WtBւ-T}r;(q7*גAdG%A*7¡e9`f8f T/r2in+ve,%s˽K?/m X|>YME3RȺsևN>HcG&K,$,|3-AX%G)FK1b(8S&cH~F9b9}tý+)&q9W5WLDCJ {EcXKɼ ˽ lCT +{@J\с9=!Au0fymK՗O S=EUб.ZG+,'޹HaeR.NLR hHQ9KT1P^.}`S"]jIALTJyd`a9e4j5.P΋fG>O9YpQr eK*f iDUˉoKdL)в`;%Q*(UgPFT/~5ЉQ ;A-k),&eT/C]Ys9+F‘ $yٍލ}`dӒZDDRvX($4*9&$F=^m?]hl5lB;jH\(B5tR71ȔXA $k}]9,d[1X lXLXZڦǓw~2Mh4%Șf%͢D2)ُ&(Gd"b)gtBԅXBN&kUSw%6M7rk&!`*pl>hX qS==ʍW3V٠ֶeO넸>HF_O M3\?`B*6DYaWcz(sΉն?Cn1i!W0|k%j+Om@V HR}mGrMoo P#pPT9ivuA `M[\o~䮝o8rCTQ{@Wߖ 9KK>yZ#;?[%{^XO.bYJ,^Iy~f.xgn)_pn{-x'ٲц^ Uwdnܓdgv ]f*u1+@ph)eq|\$Oed˚9=!vsENQ^^ÈG+oV%EO;2ӣTgw[T,`ܜrEju~]adTV}ggyuZ[ '>MExjnLjӬ'-=FVWIqo ;lvs[CWs9 JңIt۟Qdͩ\!c+'-bVLJSӒƛFP+Mق*bȮONzjZ3g0rۨ vo[$Ό!3fdK:'x/C1F)25VjͲlr2$rBuԗ/9oC5+cXg!"XvV2hydRh%S/V',)%d굤*:2;zc]v'`H3b8KXQ !8Rr$6S ƨtbJO0{h;zߞ;7Yq6;0ޙB4Huw4F.||~}tr*ޜFunZlZq>2H"vb?ݟo;fҺh=߈KFM  LP6peS ɥި.;}@yk/6]连W2*7T_,+[fʌʙ*X a8% ,` YHˢ1ʂcU*vdMͣ3z||k{V^0߸D.,Z;VoOϞ5Zr2r(DZuYZ5S$wZ )J-*?V/32f9L]ɠѮ$b >-W1Tȣw.ww ϰg ư|tfP`ՆBHhmZmpqW\)op;c0f,[(p/Dy&^&ZpJȳ2isZzjn}Yxd%. 6<*Wc: _G[PK?NjU)W#A$"x #clH_pZquq>3}k瞫)E/ZEw| nٯkkea:"hF+:ДӴ_3lR~V {Nu/ή[}\j_ dM=;eV@+(:ҕ@( 23RPt6j%&e52jn)Dʓ0gAbZ箧HmUa~iuo߫׿,WNJn㊾ :a4ȲheD-#amQ ;sEh|672(ܭPVP Dr)[bD!<[zJV˦ȪJ%cQLتTX#G`>\Q3(wd3;^e{zu㘾xsnqQVÚM T وө%0JbZY9sJF`(-d9lgd9`-=&qG;gkvTNoШmcb"(WsÊ'yPb+d 2CJ:]\u咇8ܻ]ZITx#c1+SrV* SC;Hs\ qpB󦏭!Dv-*I!F#(9?XFwDG9rw5$X҃ocRYLyby<1n$f#є&6}N9-whxGX1p?+$yLa.g ̶\#Ȇ(d%ɐFWl)fD;!ڙ|ᙏ1{JK8{.Dv-F |5,fv0e&[+V#0k+Uo*s4tvhIs,.C<[78Oءy"FM˴\9E[D(oMl»Jr= 9`ʕQkuw.|{ݾX Η#zX#ک3.G§Z Ϣ Gs)9ԩޔj=VM8H9<[4r H/g0Tӣ*e6{+5!ouZ&Ю4\ɖ#ZzG {8]4HN#ٷuqYvYz:N ye@%[> a/&;,' k+S:KֺD;A#廗GVm%6H%a $/F-X * Mr\cBX"0DFʊiTXFǾ5QTc9nk\-W ~_y(=~fMb`j pl-ؓԲxEE*&Zt)y D!6‹Vd*J^[dI֗dq_ZR怾^BHz֏tmi'( x%浻X@;OCM'e0}h{"_ <QcTHoɉ.B0ͯN`Ϲr@kV<j˚+kf]Z^XS8ӎ#zII2SxdiJm-WQ5i$^:h륄Grz#VV+eѝC4!6(*m؈Cmd$[Rm!~"UbˁRF3X^~6>tQA!4ɏ-,k^=dc@x"0aw[spXmΐx*M1 \{c2Jyh68^\zꢋ3lSwp"jYeD.KC8*ῂ: =@Nʐ]e+ $!G,X4ܿػ8]+C%uaU 0pZG=ҶۚiLXZ$.2 }ُLjiTPa4> nXru .Ѷ,e![\-Q+m8C ޱu!wz* 7}SBr1^j Wcx_V{_#ccmv%ZWuYU~ QB@`p6,2RȜX3V &D6h.i3ؙeeg$ث ,#Δ!9N'ZMT,\ hU%ijBO2۝qB_F뒻-P6Ӛe 1tc.n"dXNZ[yCyq@]k1+} 翻\IFk#ۆ ^H--Xv5js[7!;838Qn`'*}cԊmC¹i:ʾРncG AZJ5y=sd cþϮuוֹF|Eaw8{2Ntu=RfRQ:&/e e.{ha֬Dy({R=SQotSu:\)`ߺ~'cQoYάwV!.qkK%oQlwl)cin )6w뫷kjx|<1SK>b2|Ǭ41 N F@191Z&Ж 5OrR'kKk*:4(ؒ0bOilm1Ş䉞k}# ~<5fk9;_PV,ZҤ1Ԫ7B6el] c34J<`ѣB( mۂEẇT\pvHmK abdscz큩+ñYC;rꡟqyRzO!:^0([ -XhK pxzxv٘5RsuJwa:<n&Diɚ]"ǹ(*wmݨݵW1'&ó~Ъk?ђ!GqN3O6⼓JTuxDf#qolpoN[vFY!l>gmG>\,15:9S`YZ}=~3=vN7 !T+B]љ|㜌qUэӒ nz!—&ǑGA̞ţUZ3B7fCv%`aE`&BM7&'fv T>:x60Fzc]g _=fhёS+?h)ӕp]!ڭBS2^є˶ 6c*{@Mh]:1G rč+gm\f.|O%BՈ W/?jhͬdY"XU9LeՍ IR~gݏqx8aec'^vc<-KYXƲ14@g?'͡eN9ʘk ƻ^ek6$%ғ$iݍׅl.{|IԽ^ѱ+*!Zrr['T/C.WUxEJx$S])1Q_4{moPP*!jihM'kr]%bcN`ۧB^d=\s L0R6 0%SouE71:c򹭺!v賈֔Yw:ʘ ֬wfg?PZm*Zj°cD7|Rַ^J`WIX¡ %o†j#ᄈ1$Qǭ[f$&G]4?^%XwEEwxrvٻ: )@]zҭ2 `fN؇hFD5'iE**ZcHs*WC"Kze__V+pA8Q5KhZ2AgC0 !ԥigJi.g/V؉7[ؚŦ`H+TmyHi ]F7Tr/&9ɱ"M 4Kr0רU:_d tRc" g!8u¹*t$nhaw]G=r$(_wϦ4R\o)ƀ_y-ř[ZooZ]4/ꇷg#T1|P`+۲\&壳b "8"T$H@Nk iKexdcz=6ʀ!&U(~]|ס7F8,d[>+6c`J^p^"c8r P dzqV.2 *J J5@*c=[2%2WLz xI[f~g6v̮?59=D*s`EOWٞt)vUmƝM|4#يEOÐY9QE?S8 MazXSIk55L}PBn&[Mn`݋6);jUjcž5F@'NЁ#g7}g.8R7R<*@ƭљ*T ɂD].,0%ؔbd{[%89{~?.Y%L)8GE1Bc^  15 f Aj*0s)SeJIXҹt!os. u7؏[V:7koG}Du‡֛3`=(P3U²%g)b##E9 H ʚMwV˨26,_P*p* *+B"6a={ egQHk]9/7S&zI+Pӱ +`e]kv:2dm]JAÆ G@MGn}}{0+D`dN s&CF]ٯo{[cnE' sž`fzԎ' Φ Qλ]]|oL27(fPbnx{lY3CBNKy7O*V?+a }sXTbllxf4uQrty1X#ZS6юzO M"NrVW:ҩk!z9o\./0QW*ѻ=B@.Ī്J. $g㕖cr"9u(нC滱x  & O?T*)D?0D>tHLUʶQ)8"^-Xx\ |.I#S|sMi6Բs$~!%i{$U9n6=yi&(-/Hv]^&Ayݝ A=y\Ӓ{?0yYO 6& ?tbEcC-gI*7`vTOp-(,_YKWWvhG |B8:rf@39! wcH46 ៫?^>ۏ pEԛn]/OF5C]A6Ϣ:e.d!e^nI/:; BCh`peI6hPf.U?&-'0:y g^E6TPXÇ=ɶqĚN* bPM\"ޛx`umE#R>Rf@޵mqrH<̚kI$>TndK.&%JU "S+l3\ w۩T*Aee]zV^^?q,Nr,tq3a]\ ??P#YKV"SH-#.z{w6]3Psryq߷71|}<孌uG *MKIݳ+Ͻذ>}0(Gӹ 5ؒ/❩#_*";K3ѩ AjJkw!&(dpe[hS粭_ԏ^OV~wjuG{G<:`&A :9q2 yX-.3"ZahQ2 3 D;۵nrN2⼣ 5j-.Ga8#2u@U_@|yv.}9FO7kuᬇGb%ҌFCh91sGA~sS&[;3NV PK0:#cY܊8+։ڳƎ+=Rgiv|+O'oV;VIOчƀThlq+O>R{l?6pރ̙"5B<3ϙ~[0{&0_ ~(5̓`F׭&;}Ndva8TكƕDv76SpS;܋vcQ9Xa sFLj79w4U Aϫ1Xar%Vh 9s+P=qj磳EYe5`(n2"hf z$E; 5DVWfF/ Q֛|;5-9?t~?~qdrO#LpL!km4YZg6-yQR}]F:1lhe=s4xzI &X lgpFfL91,:?lf"'Bn8F?܁`<+B=9ƒgf23Yz=msln2Yt7DmI{Ӡ\Q“)Sg*+ֳߥ9?7SB?˛}`YKֺX",İ"{sI5-RRmŌ %^ s%&fkRމYv(^Pؿ?;v`o^jjU3* g"K"i[FsIGiRC;F8:XPXxM9Zm 쬇/MFōTvؗ}euwe#q̺ fYIug5+6үt1o]kCWM0[jZVloeⲿH֎h}~r{Ry+* Rw>94ّ:ihRDVc9HMFVk$1HMf O~zn PG_~o -KOAFQzI$HM$i\1wA j3%kxn [1d` , \#K#p,v@/ؖ(-5vY'ؘ`6ԟRNAVIF*q #xPd-Erb^0,ֶo,! ?x= :dhJmQa >!JbS$TS {81\`9_46ԋs߳j|*[}4JTPFL"E6PK_l­w$l0 Q4t*Ԩ!Hf!QxS@t[J/oˇ%]]Q3^_z8r I6- p*yQWu)`*t\IVDxJuW KQ${zUk`]U1F澭ւ(gw^.޼;0"S\jOHXkK9H\GR[8_3.! #, 4|TJUK/5f'2^C?Tc.|yUS? L+H1B5ƭG҆ꅈ s-HH*۬"mγکfoI}Y#Д]{ٵƁ;aXY(rw9Qi69G]ToNۙjB``~_&T b /ؘ} EH<#]բǥ(nx&Jc(^^ۨl;0(֊+@Dv?Es+81!KZ>fz{"X N!;C86Z@o"{;VU/?=P%#{4lB5tk |znD591V5n΀B?_ROp"8)yc} !q7v\9'!n܃ԽMi_x=yD"Vw0r:|q(ψ6Fv>ݘ9y7djzz1Ú䚙{n±F:ix>N::nrN1j1Tg a1OzZbпu;T +@cM*7f\9'bwݣF4 ~-}${Kb҆X d5L# ?UHU|fWl6aTx`a/ dva8E9UZJ9Nf7 yDy2'e/x{1f=ŇI&;}Jt!cvL;0Uk+༸^'h7:I&0jЗCv^*isv_>T!y,&;YrhW@; Y &{(/}̳ qaV&3.ZG'`L09jNקmnut=?}B\=yf^H* ;y]1j$> =HdU3Z7:W=JY*;-X`GoU;WH^fPH!I`*d}sNI$3 vz"+xqR;=w$jp.J6F/Ls dvg _4.)kz[`-H]$3b/Q,MFDJ +TORZt7%֨5[p3ݶPce1j%ǦcKd 5⭐1To:zGkjt9*b16ɴAdqK`cD" zMe\[E2./M~}`Eojhec擄k;Lx:E*sKq6hn u Zl)t-z"UAyftTVj1\lTk2ufp$,4CKh! ƾ8K S;{PRׯ{{)Cbx..)ceuS,,s8MDhy\=kiZԐ&մ,,%* ,sF#v O+2\11f݆!]ň0~R"t`f K׹ɹr Z@m͐VfBϵO4↋ aD(ڗul?:6 )ܗ A`z|v3AKNĬV5V}<޹B79DL=K)g(SO)1H͠IW[Z""q圪12TAS q1\F%9ax63:Įlu5L||Xrk[ph*uo9!,hsEQōGRpYܸ;Vde fqc7V479vX#l7}K?5 ~gj{ȑ_infXe `nq9\/Eh^%K%ˑ/㨩YU|T\] eoxL[KT3{ima*wm#:p\5߇EQ^\\IQpu64]p8oB3 ^HM4"ØO$]Jʰ ݉1gBl„sT,g3PGzE/u;&I_#X&-^& Jf%Tę-j*kƸH%'b:Y_T͊.ƮH_z^9`LG-^k@`3R bBvm뻗c&<>FѾm-F9nYңHF47zF>𶣺O-t-yE(KvXLNWL?-ؑ:A\`Ǵ v,b"h``ǽ).uƅ- .-dHw,{V4x96C>4\fZ>Sv/;#hfR17C/LNE|-t@:P-zu(ݡ1FNG&TYfPC}v}aFб7</>!'#FaMwXsjh|c H$m`PŢ;.C(a0mb\4>0z1aVA:OҘ3?N[!(9VךSm܋1N :/qdJ~%5$1mrQ$ wr(&bXT譟cuGE5EkEj%h ɚ聻C8IOy#\ĘüuG]}4[ 8w~zO$S/t)`*#iSAw]M#WIWNٽhFfrWssBJ@6ey=,/c{778_2_9|s~z@~|k]?Opߊ;EWEP*SsQ.;yrX Ti*oYT%wzt.?Clct_UN}mcr_8l|9- 2@*{™\z6ffT:u{8˩s""8vΪSN],kCwDG qbDz[\c흺@Ա0˰q^d,{CEF-)i %g#/.).0,J[4i#13F+^}lf8nNH`oWujR^xG>)l z%GDH·h8~up9,y̸xcT ]QzV> 8GEzx|QSiEp@h4k6|F2:5F0Gđ'H*!m>+yDIHmB !; ΁s9BVB"2nJFrº$#6u"4"EZ-Xi"^u-:lls0"s|ѓQyxsAZ:5F<y!y18$N];!Qݾ0;QG}954|;,V+k7+;Lj#=<<aXk)k>b%[3_/@JH3TD.p0)(ka%Y|Cc8wW?;W"DwS=7$bcP2'WɟޜĮB,\ME[R1'Zs_Mgp&H2_AY*[FQQcʓtD|ӈ!U`^$CĚ^ Xev{BGev0&"5.1&cJ펍Ʉ8 y0f0A([}٘nOM :sڅʔJyo)㛍;tj'jDcvyh3bJص_$@ݾ;QG!Y 20lZWR] EJEs,zn d3A"MLv${_Ǔ/״}ƽ="&ꑆ9Ҥx6C,ZzAݣ<ٍ$u=58# y8 żH{lZt_3t ^Fü HV!!NAS6RGDMy4sAGϵB1QA4'[][0/ZAZRVkq&3lfIgxy{]5Nʫh_mfoiBгRD!gx\2[\t-:L[Ym2^6HߩA'g'qkƚuLPkl1t.if]4UZXҀP`dnWdupc砺"*I C}6,\W_.`q8Aʫ2zr}yySTiu}EַyT-vqo޾{xGDoy{\ Ls/?Y$wb^mh|T|.grO:9P0EemP̡hb!'g ӻ~qy~vݚk왺UjD:\۫fK׸2:Ě.Ox?+g6WN* @OgidO{X/Woc\J{ָ6ԞLmcru] ϶b\ ʛieKBzB>.ʃh"]R>BUI*vtp_I@h'SsPbBA]pd`ei8II9|c3 YgW ZnД2dr{Ol1"g鈞m<ve cDzՈw84NІ?Ǜ8욥\u"Vx[Oot`QzDF^f'Cާzen.ѽ}:j^%YLfzD)K47M2f1<f+nXnOȻ`i"X=.oȈ879"jgOcPnRfbR;v]?xy}_K[3(3ׯӕޔӻO~*?7&̀L`RH`<2y6_Gw~_%_Ee͎;mko_ވd, xOF('D##G(2As42 ŢG|j; ZH], ٙCWZc6U*e.z옝5;wH10eN9[%}C%,5_kHwWWe8ݛ_W?[~cE;nᣔvKy'/gM1(z}-?9|Nq?9*O+"MTcΟ{|z6Xf x-=cg_ˋ[i\%M@]OϯO?kZiˈ&v[&Kw = ъvώ|n,CTv͂ZޭZ|ٕחlv8q;gӿv1` K-7MqhRX"E JLE-T# IG3FRjG%8L^~jƥ6}BZ^mDf>L;!4_e\-zUwmm#YzٝݡS? blAdUV,[ZIN4)FIELG-RdN{^zs3vN˞_v/I W_hFI0|S4ְނ9 L`xQcvJI\"'5N0ssO*Ejѡge8&(cMiȂ斂ec}_+ 璠(:EaaG0'6QD"i+$&,J4')Viy\Z椰I+e3wz&$) Ky2RQ~_ќJQtc?1%"N-*@Xo:3$.{}vrle3/>8` lnQwb_i]zXY@ԏ=l6E%Y FOy@ (i S6NK IܲYs;C, bP-Ub#b.7r.XD#kAv*b<Ҍ"2pj&H8R9{3Ւq s ,T{u~<7 g|!/֭$>/*vMhv`Ȥ7M)p;{rB$AH$V["cjC HV u94D\Xt\ $>ѳV"h)A+*fH~vYcJDMH2pV-cDsPlENh\Ÿ""sZe QRw V֘c`B׮ŬRs%ۺG Ӓ7`?dyk]hL KκE_κoֹ|<L/-2r½}PK?t֨X).˫vK5HDQɬX_o  ~ugF8 ' ɏSNnw4yeN=c2l~JLUtϴaZ3-P9O,[K9Zݔ`K W,,C'JQ/}zXF}JΚ/rT>(QF81k7I$J Rіwfl_=a| T.i|mA0])(u:]{PvAgloTn.`11ԃL"qfp6*!ZPCR ;4w(.o_ϰ5YC>:еr+qdip#ӗQhךQrjdH+\5jm T͊@ryQs=M30g"aG]6[78iӷۈTL pȚ8o9;`JZf?WᶌjԋmK/2_nAhĒRVQ)"j%QoX&;UF̢GPm-X<RQW9٨|"]k_'@IǷWf9!:è>ji}/,uh=rѷ۟Ay׻}4Q?TIO^SN Tf1EZjO>5ś{fw%G 辍L.(װ ;^s*GMْhJC*r~/#?DXpR$"P%E aRіw1(-B(C%3[= [lH,LH)K)X% ̋  k6tأ#IS)+<=< =S;XPW0*"xV]y~`~E~"O~H!QA]Ȧ_XQ!Nبhua!"lF13!;guO2a$zg\іa/-9 ѩm(Y 8"1&I/XLI ] ?H`n1dgP:\i#~\!AcBië-h# /j/c4Uλ0Y6`'#2 `0_ll;sHt|S/P6R^\"W]z]gxʯ{PEwSY%큺'Vw>(]짃yN?O~=Ʒ3Q4pl豄5E3ARZn狏3D|$GKWj7هGhA( d\xN- CuW^`X B Ɔk hmAk8n6hrNAPEPc#x_iFQ[,'|n(zeZ~_¹3LfB%ĚqΟW?r:B}}e?qSMn$'MkMf$C%s)N2 Sk  Z9i K57MM,GFR$gXR*pg!S_SI{LT}Q5ywbxd;.{,d̸YRMV~%W8=Ihְ#ha"8dW4]AK  aKB.TI@ל`–N(liڬix1քHPvn׾۵ov۹]۹%(Br3iB9+b"1ICTaDIA M!$KKPU_TIW3& kT(*SaN{q\A͛ lvy?xR?i7o0:#X3Lc,g.^L̰~yfЙ5d݀Wafչ@?,RFs˯lhKÈDT]%KA@%V % ԙ⍺[:gMX- K׭+4 I+ p>BXlJHXV*sO=?ÑDyOVQWR|)1TÇe>Գ[O3Vз$)=SGkI,2J!8J%6ZrId5/58 ZX0 @N/rRbР4e1!x$&r)#Rk`q vFXKt8cp0W!'&?tX|֑H1 U)BS[$X)W5v vƮR\$2X) FR++RrR*1"Ǒ@2X)a}dTd=UZ$ To)֑aj:|I$E$`gU92n&>"; c,E}jnX`RQΊ'%,&syfX0~x0JW?[  1m9XsS}Sw- rl4RPs9Y@:A|:'|/\r:־᪈iD0Šss;+Q+LL^DzSeG(DaVP8e$Wm|gN+˹M DwfDUcaw;&\Z"X!>\xczil&(_<{80v}hTۡ 0PgI{8+L` 61b'BR}ݢdSjI ;l*ޫw;R,F@CFegvj SYRU4cc>KV-~H%P ͜Oc4#RepZmj°( :ͮMk\h/)(T7^_N.pm tgOv: A Mfk8ah&e)!ڛ ι1\xe&T1LsZZy̽ݛrDI\y=bqjAEsQ*RT~<%z̀).7Lxigh%^z/+iқ|lRT M:l_êDߺ>_!b0:SF! e5DqfU1CuROIav@4QWYc *:Oh*g Mf~*E- 4d:Z5Bõ*z?,hz4e)o)_jE|Q\-KILh@ʰVu#Fqz:"yZ0 VpVd46.KvVe).hj6ؘJOV{n5C0Sr,tFzaig.%j'W34E}o)>"@rL}Fo\?^OOE^"7); yςr *9M]Ha6LOHD U|ł=D#G}LS:@W:λ<<fr-CcLa8F2-|fۂijL}f]ڝf!yFް!ʲ+56&rSõ{_9xb3ZWZiomv'0Lźo+97q &WI ?WbG70Dɓr81Wi(B*9gN9>ڶ-ۅ*Տ>8}\pv\pLJž0h#t#G}߇bNqF@稼wutM u(C.̫:L&n`B,IN_~=k\ R49uwx Fp]T05Ɲ$RmXNIERY 8kZ,RR:"cvY V,F.~տ Ɠ&[%̝;@~/'bw@~rj7~P ?\Qm7<4ݓ,j7pPYC-(RV@?o @=wyl2}7Iz:j͢w'\$A3=¹ԵW/?MtF.ĉyGkdnSR]Ҷ\rޠDXkׯHqHt%#NE $bE]\~CRx}* ^%Tԡy͈QPMн2%Fv'o^Ee Yas25ՊqF`]ju竩oTUjtT[NOIpbLupثs7Rv$-0| @Χy[a-dTryŎ+]QUC*8"S ubЩڇR`̰SR)X ET,eڛY&TLy*Hp΍H|.+%a2s߂}_yj𥝡ٗ+..JN'0ڶjePvW7Cr3iALei,9ٟZ,T# X&)snws 2fPcEкg*v 8(ItZ' Lٹ8735Pen<`alzh~3!}hG=3$GL cBe[{neqTh|la/u6բggOv: A A1lv!/Q`sC1!;N7]y 7Dcn<f4,VBFӹ&0|Pl3 >=:acK9UMqYms! 0k6 CuXsn`oC䲝 [[ǂuNPK+EaVQ}7>[oFupz2LVP0j­թA0w$!FDK*h0tC[QV IdLڸ礵N˷]ˏ?_{^, /n4 ~aѶ~ㇸ]lrLo'GG!KF&$rZ8qkF2L c`y‰6+AEmhth60dv׳'73[N籐^M8KI`noO c2pP<;} Zavo1!m'vvs|៫GvRhicE(c{Gſ^R(9DxC6e7S8i6 ֋zX/RE5MwK4pa IyrɁ]`#Vb!(Q{9EzLy9z {R>@kEs#f[/xk=yl t(j`&opJB{cJ(6:z~WW[bg^^^|m g􊩿_<r!P,E)O|u[l1V1#%0å`-f8LRF2,q ~ uI15"FE͏&? Ph0ǘaUZiO^,rZfKYF)`.e rQr޶.ŚeD޵\|N'0bWOq~]& 6wH-A=4[m߸Anq `{P݈~rbEt\{W\- `/n;eMGWܚ gQ9/@E<3H rOޢ;+]N #(}b\ }4H 3ɃؐLuG8#?G:oNǟH b3h9Y{F(h}V&Q7^_N'*Tj5z"LO/j"&~6h2à\!߮h}{LDѮhhU4q;:P+(Ql8}N[s7(+N)'uD҃|T4JKi*bĸ ]kh+Og1"8U@{31=; ?\A{ '/U0kppO~5yJ_܃%Bwʱ]ݬL,Ċ\oBzhm[ɷ+|m`3XBqqCyfWc/f>fju Gcs ARicU( umXcs ԛ,ږ73xai澜gŏZ7 Bd5SzOQЏCcʨfdh./νeiʞpNQoqOşLyf>-P;m +u*Z„~;P܇B{b|rJ:XSڎ] w{"LSiMU}{? B]1Qb'5-AweI)2C@ڽڰa@S,ʤpYb..$&2Ȍa":Kd8#&2e\˔QrY(2H}Xcfƕ1rkK8:Fŕ3TlF/U M=sDi.p1t =a☯B^:̺$}9#ŜIϤRW!0焓Xqp)#@TWy~|ר R~wm`Bd%ue֢0{"_m-NkcErHkZ$`˩0uI!6N! @ޕR)GPnkQtǴ3HURC"w{*5;NATX!B[G apE"mβ3:,8jHr2C@ h=Er" (VgSˊ c3tPs# jaN #X 2@A 뵈 /`308`i()ϻ)[?RH(JDlXGAuH)H0ˬgD "c "^s$u:}Xcsʑ@`\a5< (:F͂E3)ec83+Zq_k! cƺ3&$>%ĮJXStS Ai ˂J xw@ %DFh4Y Q@a tXSLhI,SE3.N",LH mjBykx@BE hmL =E$#Nbp&bg(;pgv=股d\xs#wnt+qs.mѸNi=yLw|-XQ  ޑw=4x4YOy‡UR_^<[{eH%#eh?.߹1ܘ,6F:ӷt`h*H(Xq&qFZ" <YekT${B=Į)J XGkm \ziSO ">Q x6`#4^cn) 2Half>7!>/y6+DCf@ĚqR[> ޓfy"Q+b\c6rYIk2BM1S9 $9>uBK$#wa켑v^ Dv5}/ݪKpB&N\zK9[bi#8M)hT&6ANu]6\k墺mN[.B'8np"&aF->ON ~- qr2>=Vڥ Kk)+dbnnvKw[@nm0Q+n!62K33[-EL$ ќJUqv}[.QU[,thǪ-/spBƔL9Wˊě6a۴kn}H|Ssv;6URz "3u.“٠ӿ{[HL)ū.߯V:\.Q& ,_$J$w#HE$;]uL&K+g;xͳNg=)ի62j'#.%SΎi%&(8S_粮%Db ³kyܱ  A7N#M=jƂV٨l'_&leU_RYV%+Mi ʑ|tt8< l>/G5 2F=d)cκsE+ǎpTJqq}6+Nk3=l\\a'=pQ*/|8wv2 "Ι~fa`ac#5.]e?4*֙{>^-n`L[<(0,h #QHܻIf-?@\v>o[NXrXp Z\xХ}7B@dh Q C/ʂUY\( 6 1L] 7秴K9lvmjSŽ-ܗ/s`lZGS^vfhg 9+wSŧH O 1`uUK[(U$JrT;J#~Oa'Ok2U⿶HR]vh'ʮvV?jY󈔝D 9SΣ\`cJBRG^\HR>7fލ~u]D7Xkn봎4҃jq]v֘W̡}ݦ 4]qT؉rA sl?9l1oCWf .ZaF~7ҠL p)S;lJʋ,*0>Ss5™YM)Nve~gdG<̛۬'5Fʉy)zڣ\ xͼǢ?[Q9~.P۔ty:WZr)j/ 9Oz(A(õwꩶIc(Mpv:Duf0EfxEBZ'1i24SG J=@4M1@FpnNlM!} }V5k cWW/zA6nKDz=c]XfnS-blǒjdJQ KjtTO}}\(LNB,sfT399+y :^rMЇk9_\5k^bn *!1^ou Ups]AةċOKߛB;զޭ™ sZ*=oea2o>~yƇ7}qx 9q&BzrzlqzQ=IvpV&?[Rߍ^xKB5DJzs2[g)r ztbz.~[]/MR|#,9""QIhoy>cΌg3,M*fofWsV9F0%𷾓),kI#EaURz,fkeAqt;;ߒKì;֚I>ڰOJG{ !/G ;X`d`.P 01񰗐mT* 9sfddt8K]?JNz-{O$FV~6ߔ͛&2F.w#&%s88'9Lս箃q|:! VS;  z`0[?=tLکcy}TևبjZ0m#1U%H!iӨ;U#AgZ,o`՟ڋDq)ԉI1p(R: oQ 6h}0(!plAO>E#qj LbL#DFG~8bQ{kI$Yĉi=2]tHQ/+^ZBt Ô1vM.׎ɪHGooI5>M˻:nNq~j,g7s;ۼzN/fwɸ0$qWX܋b.mZ-Q:UDylm($V|" 4BBj,CqBP# 1鈁pt!@$i]eCrPJ^kOf^3y[O!dy 4i'V4NoLxAք>Y>- ?7W n5{[I\>^פNrD2yVYJJLYOIJf~\%IU*ji 2nΎB:zy^ S-EVm>ݐ g)^(Gxܵh~q-X\B*ZI"k 퇠UxČ pqަdp3w_/my4[ N5[;k"OSmzKye0\WDL}Mn*j&Sߋ?t5Mu *"G;O(8w~|(LDu=Tz~.UChFNi]~JS_kk,+A uh-ȹ~+ .4̋Ysyq@ƫ.`Ce:Kp#Ƌ5V:;F!\I-KI6^'$zwwKgwNP2vU?Rsu]'BYy]%dn7@I T֫s[Q(T3Tp[gA 1 9}(eԺV`$J8` 3p-DX-+E4%r"wD%F9&!#a)لc)$F;*^6H6LZ6N,kT^̑gkϖkt6VC L j&CRנljc-P1Op[ӛ{_aRbv>:@IczsF#R(h`Y(dI[7J!pGQw|yŠ mcH/u 4fiFߴ瓜ikhl'J $.eX )%t/$н컮LV4`h 3v60/hYN/_*Pa}tH08v@uziIӐM*`Ce3$~~{ݑ?ϊ>t3ѡwφqw9ߨ5>权Wi_tW}?t 1wq!||m3bԞ~ s99\vƳ{'^4x\NƷL5(|)n>u[iϨ*L/w9O[N&mt5uɭe=q2~,b+7\q8xxuSNt}gx&m.4>:M]:IM?N/} fg^gi:y }vg+3N}!?]yn&Cͳ&0NN(^ ; mӻr:pe=!,?]OI1A]z ̄ug 髎GfxqW_ARҤ<* /Ε^1<^kߌ"9 ?}o ^ɄM:a/XF Mm>:b:lt677`\ 33ydApg}3o,72=2iAp@~AD'"6`@$1J"Qs۩)7[;0@1aWšB @;s03(rh}/8o>v"ak^J),w뻽9e ?wsXۥno3dU/<]'~mUmΉ|NT)JbzCLI͚%iqGMsv81O ̔hL 4I}~ I Rݻwk`wƖ,CH G`ٔ XN$ϲUoIѐlYh1щ`!#vbRj„ŖHiNrJv% 8W *%h^kyDϲ5Ѽ&W./<0Z$N@Kf#!q% NcP`#ilyw[),`~ ~J0k:Ђœ>F}9u2fc7},&Ic7}cѯ9Xjpi9o7 (KBL=sZƜ_bB.眂,\;KfTƒ$SޙY>2Op)q?2¬?$)0 )_B0zZyhV\#kj0VyhwN W:mAq% $@F1 Yab+x V%L8#,ilI~8H1DڝOڃӂIV()?6hxv%xUljÿW`OMՏZ [Au0on:ѷo77$ xj_W.`g޻ޞi{Ǐk}1I.$huėզ8k[dy8!x*N;5DC[zj`Ds*5VAzxh OT$DƑluMy1KL^ߞVaqډU؏iLrxWͬvZUhk|)xz}- 90 6ײawyM 0hSH8` WJYD"$bB?k)+h? huҔY hR &W(aSB <` ŚFF8m^ 3*a940WN79V]TOZp>|34 E-ʊZCl1)&5jV}.HvHӻa4j.iyF\ak 3׵~/@ibw<ϕ+AGrV{8 5Lu>2XI%bGJ5Up 6  3>^^xzatk0ϻ"Ss/WΐB!MC"Anpq:C*B}o-Q!x^FiEQV~>h`Wɜ]~+{ίOӽ{J۱ b=PZ=g]uVK ΰC#Vqvw̛.r,ٺـ`K~K{q9lK̶>6a٬;'nE)*mY̶]MUehԸ*8\TgfxK$LvQWҝ7|Z/Oﵲ=Q] 0Ӱ |pՍ@1\O   !8?^xv`&`_iu]*ѿ:ʙ*-iciDorg3.ZƭcFh |11$gH3-b6,Z4iK!a`RJE aD3(u<\WU6V[I|Iڭ7iA˕6L8 _(-4 {v/_G1J蒩dPs5puB>TdqV8|ӓlmvëBj^ A<]򆛍H*sg,ЈfZXlI(qBV×db4TaV&wBB$  XEfc՗0Y-f?:9vr_^`dxҁQR08ZO~dBŒ QI)#b#'ZIiL*B#ˆDBXY*ƀ t[9)1G+odDo55m\'r,pr,pr,pr,(1\J*r"D0DS`LP*R 2֒ڈkvkxhh$C,r.BdZH B(fb A8!"K7E~1nqWf^t,WP"m$1S&)C-( Ilh`0DG$ IMm5*ƾGh}m/vMׂbؠnۃ.7{svI<4`H-9wg 6E|_RWDz+ZjǿݕZ!Y\8cy_ Z"魼a+PW넇3ˆ=Zs &|N~"1BIDu!R66g !!Vl/ti^T}N&"AuѶ5->]/KɭzϜM ;A}xXR4(@5_:8#&@ZzQW9YDG[B(pLY,`VI b 4&2IJpihlII1(Is@0T&☀ )( "Sm͸5Q0 ΆFo*@'Z8aqAIJƘP{ L &8RqhTay4A[ejQx^9tU}OÔٜnWm4)99爥+b2mpL_L[duyBp-m LA/sj;8Wz^o v'asL_g]-%I$}uy/0~*]=%ϔR!pg18duU2"ilUlVlSbAB(K %qYa lu}D1ɨT+اТzHR;%ggjƢh;d62&uk(_DLDd ; [$8Esz4Mq\LH3D,= JxwϵO+XI3 ~@Ί;76kDۿ̅Z^,ihÅ\NޛTO]RH{zY9\86m'E"+ŦTX9jI=1vz2&h@dݵv!KfݔAPJ&櫅@v{ {L ]whz$\ލMe,E׵[VRVY6LU-kڭ E$bli;u! N!n]pv呔A{=io#GvEm^8v İכ/k ꜑i%;+[MRf7x[agpP{*cuM^=bʯ qI/_f>cCPz\?zuW?K#x~F_ `rjY`Zf [Uѩ&Z1`E=cQ($0'=i0(Z᧪|޾s]VZyWص[jEE]EZc0y?\ZR$o,O^://Wv#ҍebډT0'u?U,"LQ,T׌׋t҇ :ܣ>-v>hY7YtoҖW=rO]TORAQ|yn&b!.&X{Ʌm˱q2J)ns2&T:"FRkIcFOf_.B2XFqM][_18~/ܺLT\w ~sݪp6C`$Щd-V!7Ʊor7[uw.7WI5U-P} ŏ~hF1}3"n?>4fI4L|}?D9ݵ>_{|&n`K^n1(/?cPڢj7Q={< WV{SkoהekNH&m>e{4TqKzBQ[^GN7zȞ[:&sY?SE#Ҷ]{6RW/v`x܀_my`V1u(}4$ Eo44Ϳ~SQD(}eyz"ǀ@4SjIK~4dc2&Gq:,LiцMH@k?Tϯ޸w)6d.` 8΅`\SxUu(gwqY.\jRSC婁5Y(@'[1[[ k7ԭ{t.ޕ_u-mc0t}E|lwFg爡0eQ詊$Q'51qeNGU,C29G4m16Ẅ́$j"-U\Txaʛ3Ŵ)[nr}UWjsLˬm/$2~l=.ÔOrSŻ3ǜ˓h>FKeD T0: \"`L$r9zCAӎ<}ack,/9@' MK.'dPlF)ZX#+XXZǺ(GbJT֘Fm5ޜghGo,GeeT)rI#/ي UKqY% MGz+Œ0DqVTYFmn0e c}/p2!\}#"9'S(^S*]w>zldeFvrrxq1@(rU_X&5MeH^5q}ѲA*Fl^pyԹ j|4ScVMc3Fnk p d&d8/lJT0 1s-5!p>D{t8>By{[;o3Q4&ixfºHzK FUt"'OTyJA`gYu3B_z}hK$r*zH(0]tjO?;5ʵ8 3^/ҳ}MCs'Z`p x.^/W6o?a4svʾ3~bG0(Uʣ?TIjjlJ:|zr*n?Xٰzjat4䯣vҲT`zԚУg3'uzf3!մ/r&q=C&wX+ J=yUU+1p6c%& .-5BMXƽZ󎖳Z [ɬм%:E nU!xguhªܽ~YnեKW.^Ó[e5J:Z>=9cpEAkkptw}R18]oFCqbz`@>( C?L/a9CKC.IUHN#&" mήQ (\Ryxxny* õVHP˒UV(  1=iE>CŘ/9X^ŕGˢ Kb$)`LG.Jw@}^N=vS'k(ޚ9hRaԖ5ـhd8Լq6g}ф:9! zJ7Ąc:ƲB\8i%H(C[?jhw[kPF [+ pQ>F#W@$uMӂkIG)t&OE<9 iW.ql4'[. 2z2y!C< erL*%8rc]󓤀ykr~^-caLqE,3y4N;gECG2jiG6daXŕL9̔@%:db@Rx(k *c>2S3RL{R?#{]fᆮEJDs)*5PbXQp$K,$8ߊq%_d&`s"W+'sE*7,dT22Tz Oncrj%Ax!1HH!4n j)6{6Z ;6XA6"3 ɥؘ6gV=`Xl%16vN"Ԇ7oKny)=*}x2&yo^JI"<{k)!{eqRSJՃv>\"3vXz3=b7ߩ'O! Mk{6Se?, Q<[@Ж<})H5޵5q#˩H_\凔lmu˦TD.I9vR3!9ghfSъa5ݸ| Y ^0+2Z4!zNC  @2E,k|ݖɄϼ!mn|\~oz_ e˛U޼Mg~pm$닫Oj~5Gvʫ,tj46EJ H_%c%yA@X1pW"-mƠܓWVH9b= YABz5@ hddd9[!";$wQwyTԴ^BGQړW9gH!~E( =mq 菞 9I)mͧ'k2% ɡTJ ڤT#P*M8Q *{Jo8$K%fj‹0P0"UbUbi sje+Y0h AbmbKҖ) c:vM۔ؕHܮ]EpccJ&hADBSxxKؤ$VM7BJaRS1L )mNBZD4?f S`$|:pTS۱kwB(`Tpj1?7<ؕ1kq[6VIGBU! ><)pL7 bu"Zj,Ĉc Ӄ_㝨u1v7S*7oVp?/ ޭzw:W\uVg Qk974i)U{ k)snCzM K? YC4~V@j`%P Ē!!. Dp'quʹ(#u C,"H =lZCL!nH|ʹyP *P aj" +:Q0 uxht.oT[{J X[j ("WUZ8$|E.}؁moƱ[gir3c3yQ(ef.RM,}F2>ٺYihʢ-7TB(.VeW[em7`}|Xo:3=,Tc: SI Ps!`ahXpa7å{)J!27r%_r Dbhi.>$PV,$q/ae$C<52zuD&G8 &DBh%/F+鰐 }\1.usE;aJ`2(a0B#"2d9J I|.qDB%'`. ].vdHx 6h"87"Κޮ$@[F'buMYL:zX.b/cWOFSK5A:ArΏ4x :YHiF̜Oc.v&+cѵ1qlC~)q ӬBq7I%"I '-ns\.SbtΝTyBh|D5Ē>⇛TI!9Xyb\:C*F55qrZJ򳝏ݗ7}\dLEN $&Ω0 9mYv⬲ l/U^ )0rݝ@N 4P'$<̱$/.-Rc^̮ڽ$飼~vq8aJd}MSR!5/xUp:A: `_جDm>I?+vaD!|Zp^=t=ktNr]SB]6[`>'0mavȃB%pw{Q:uε&h sb*:gq  '<{-q>[uN ­,(p.pPH񺠜Hc802 .q9/UAh.휵Hs11?Z;/J@ddj%sK2E0,'qxOO3r2lCԻ -X.C䯣x2)22M-횑wُ^,D\Ln~q؃_$):2):2Sƺ3OkƊ[)bHigHv@ J+f1F~ɤ jn>ˍ-Q^2Nň2<e^xy;/߱\%"rBYJS7m?3:5"g_xX>]&*tzwp?^~IVtHe_4@rq\{}῅IWſxq[ok>`Y2KףE\R=_ƒ6k`KꦾP$>(P 0T!e#.9tt8d۷d岊7E5)lJ |H0V#ocnJ:_UB{UkPoFA%R-ۛGR}c*. i2ydl{X;cێLxK?Z5;|1Kv;_Z^=.r~p3_}~@⋿V-}&8bmbxȕ?9xэV" `BlX>9~DBA/-kPt=e:Pe빺M`@I(ޚx[ayq^Mߓ)>$Q_z{,AFj[gyEgl?oct&O!Ri|Ls{SFX{  E}ܮwd}HJB.51R8*b pV%ͼ瘣h}H!!}^l ) 9ƟE}(!u d9*Q7%uӨNr:眩 sTc7+ EA 'E"XNzZ]4+/)p# U8p /usV Y8ysVfJ1|%8 Z} pVE✘+1 YE #[ݞB䯣64kɺǻGFٻ'm$j(Pn+\I\qpr h뢑=u_FkD "ʼn'e{4uht(|>Ѯ7 1N3x2rGEW_$VeqfY$ jR9JE91U1)1JsGKR(ƣxJ^;@_ѷf^!s;w7tA6۩EOn7ax~]-m9sQ>سpV;S8l뜎yVt_BU~o6œzr|Ё`_?Dj>5 w#yːC3vQÈa//Q:vvqФ0l9 {n-?&naO/V|SDѱz,Q`?3ܾ:b⯟\fnl翁( ^M 'dp[Oۿz|j+Xtס`>uw/;#~6s:AVl gߗߧ`𻒌- ;Dc{C茊)SXlRɬ̤̰LiƭVc'8QNPΝH%6/3*z Mad0ەz%o&wS`ud~?*/hރ6~a2N {I! {ws|-Uٝ:=xPo_ ʷۯg?ko51b]s-@[4uhj@TA/U=7fև~:oO%IЧ=6hEvJnYLHVjHB7DvE>-1lkZ"iQ"}E~ޓOBg"uZUks  lZN2.@F3'tY)@%Nn-lt2xB\fɦD8/ JĮq).H ˴a;t;Ml _Am (ԗK@eaCJAaYSXUBr%``-o\ĺжըhu%<m m['בK.,%\e]yެeVF %e{ku2MU+m,c8r_x0Y57R7,Gia M0%Η/=[Xcl,ρy:?a̖#h*@Ie1 xLI8R&F Y2DB§㆏e2S,+NeZ de L%72MTETOMq{¬( Q8`3Hq1nf^ KqB0S̀ZPLci1mJE{^-M9h@]wqD-lg7Q$78-7~8q W :Vlxe=yB`h$fs'E\w\P__NFC>>¢I.G|G=P $ĭYϛ+jFM\4ld(MA3v`vuœQP\ {>vMg+^~ V ltrޙ)BˍƶTk ڕĩ'qW}b۝=ᬁZPIGBoK&3I75*_Tt蹕b '֚g Oh.A!(UN8qj.$IyȲ'[-7k!l*VZDb&+%\0e 0$Mj}o^M'7_ō[ME=&kwUwS6_G0B7KC'xkUH,s74Xkb&t+DHɌFqf;0-oE1N"P4CiBATZ klMo _?lыik• ñzHL/-"~9L9׼&5q.ڟIʟENXU|f.?L9UlU-m XEk=>4_77џ_<+1yyRԝMPr^$^T~Y)Pc 0t'Ecr.ЪR| kA ,k^ i|28&WK.2l\a4|^2~ ń^L>Z׶ѼH?._,zlJ+sŏ]j?E/ؙVYWon _h OT}.{3k#)fVQnh=ŗ͝r6|F7:/rR*viϱ>}O"?M|V]0kn~a|]݊nYaMx4ڔn98Ty P FE_;y-v<o{a h{N@vy07vG240'>13N44j]Y[|t=<֗f?,?:.+!PVo.sR$+ kK۬U#Yg/>0#%?'|MNfz,wM&tٳ!=,>t^SYPT&b.hhRsѺjfp%/ttZ~ep. $ZTDZYl hne6$Ff0ZdhؕsKV൲RAvܚ1P U[[|jFa ,ɒ;J*C/j&(6m6ⲑɧ9LS(`=E?BO@8͸HSrT38F!3;Zm$+p1|`֗d8T8tG +5|yI4HqJMY.?IkVHϥ'ϳ=N7ʺYjkǺ%gek"Ad_oK#EtM@ثzK}SECNS"0ǣTIDwq'w^7w7x"ro|ڎpeA g.bknuz&^FNkbQsQ_->>g ޼Ǫ5n0bvXqMpJ_jgAXi}Ym`4$OڌYYML7Ă)9ItX헌Of_1Ov0U,#nL WK{}b8 98ʁNoeIUb}P ,YY\y+zxKXl)sr϶n$$ꀗSD-e-L eGz۶ԝ?ͼ͵^d5=$*_vsJP us/I{*$FR}̰5Tl3:K*T4kš(R *8 fJ]ð9 r.mVzPCfCHPrfmG@mI/WVQ@g]D܂w%aPH4Ao&jnz51RNo^a .?r뗼`?EJGB樮ƏLYsK[*~t ɠtu8j8,v!~/*QzKV0K(xocNdo:Up $ ɺOO^l>?ͦH Js~eżoٞMpc9[1ןHgWZ5/B_ʱБZ7ԗbR:$ل(JhizZ9bf~Bq#2__}AHsE R$twPe 5-ObE 6 dбM]輹0l*7?8 ciH_~u&XĖZ Xo B6@\:MWU)SPD#lxJV{}pHk㈼ׇZ_i?<Dնw,P1َV+3r<4pA-ZlSUqbZW23Gk^R09Uh'0$)N,=_Yas 9n$ d6#tM)!t_{F"4 EΉ &BWk.[*/bzĈr4־x*G$0im"GIQ Fk.vtݲ>lm$ ҋ쿠 u(krn1)ʔwM"W (ԧ C,m4;,*W1r ~ '=p m+rVlKҁ"!'8We%1+ʐ\-J$c?/oǚ6@,$DUT"`]#E؁`(|Q;f/бfiHt(kR!n)8T ƶP}jԞ TIuF) iP`C5 o褉XB*_ B#eE{[6̡{UDUkP6~\77ҞGҁ)3\M\xܞ΄@6o^]uQAY@̑ۦ{3',1|cA^*M#Rq|5)Q)!>d/fX>KNFtUHe9iai7>Z(eUp~e-ԕ3.R_\Rou yoK.Hv$_ RM# r _GIQ(5ի+TCh!C>dԟd%CDm*;kӅ:>_ (؞ rc&̊]QKz`a])ݽdip34}C pz?{78̗w =̊IH`sNڸl@\B1}'=ttdWlRTg$se P8zl 蠂r$"&n&;Q;w;m#UV3X#(cHFPOW:TO p8>gb~ s`}H1Rwݘw9WFDt?QHN dpO.0vj߃BKa~xw5񯿆7tyxf~-,x_}cQPz \&e~z=˭ki_繏$ȿ~?+⌑|DoH/D{H=h"WW^0`'iuL_ջ 3,N|,gWF`a6;^QGH[pgOǿqbz06s jAC%<^gbX= ʹq,6oS*#1%tmmӂG2V%2Qnnx跥wj^&ʋ$~"xHGcиɲ|Yl̪ ە@V2>EJ&YtDفXՋ!c9BB{S*K/k"4cԝ:i4-B|W[ᇟ?Rr**X"E, cyAS\h,}COvdځvhxw Fa^(A7T}=Tx#U湖eZbOh^nLq|"XGn^.+xyH5I+W)F=}MhhȫNn?iVp%9 7ewpR̭)6‚PRpKc' y*w%Lqi /}8y7zwʽ{|;fc]{` t1 шg"Ȉވ( բ=XsXIɾdW6pU`}0p̈pqt}S>j@Q-t6/˟9P\7g8b07f}8^᩽()c'<C?SlCPBK@ΡkVpP6Eqj:be <=s".ZiT8kO!P"Rf0FJӬ@eh֞-8A'4*; oQ +[$\ >vDWq1ĺO.'md>/|HhŶYV~tv7@n-c1{G$4B>󛫕)93CQ| j#>ampFwn˫ᘄbcD|6 (D@B.S?H(hȔS񶎓]1TzJ+3U9Qe=ϫ=U o&B=0j4qkhY&<cyZ\5c}G(Sb(shGt PWJRuB]#J;-15&t2Ml*RH̨aK.ػ޶W}9(N%b-sb4Mϗ,!JrE;KBI XZ;)&?%,#=ۉ7j+xn۹v5N*@ȟˏ?c9m9<1o;{@``[gfVffnY2_.j!4]=-aIw{$Pԟvƭ {9pԚ6fgAr.hd<"aAc˟7e85:  .%]6N6 TaP#5ӆH%7u0.AFH`% _nl߃fȉ eo*-|i33 1J'qGIBi@W"=N8ֆQ +XQk'gdBp}5\ j~ٴ0WfE*P1%U>Y&Ӊ"F͏pO'kARݯj0l0׺}k3>?c$tk+7ۛ–&K4G϶[rYv$R}Qaݒ]zzwBOƄ|c%PR/?GuO'0,I`9;B'ʸԹn6ʒ6x,*a蹸9ͻ-z|}sNP=5zǬ*&WؐL&+uq yoߌ?f(xp;}vc 2v=I[0}LkF}z&PGe> x>P /6N>z(-[?$&v%a `e@ o'2̟%O&umVmBP( my[Ӧc$19}g^Tb"eSF {m.l6=2J,wA#qr,ײ`⼰r7räsujvϵ+M^L-8aC(CƊݿ&i05</Bk&:=۬DqycDBb3z WÃI2t $ QH$H4EQ)}-q Q1O"4))W 6JDuabwi&eؘPʲrQ}ZpQLO 3?VWN~X R*` *zHf`5rOK{b9'pa2ϝ>؋MKu̚G\²I4*֑ң(B RŒ$6-ӆɑT>y=_i Hv>\1ᒬ̴*}+o~9]tJzQ]Dşe FPM[,|#XHD8iPm&GVPX9 hb!C":bU ,#A]G 7@& V G!לFQNnN #/ʾ ΢$2R#-3b8UUne̞{FE#V6>%Fi ~[uEvnw/zZn_D˰:)Ceuz?-pwXϻTeJ}V>n: cyuWIkpo\5^\*Öө "ЪsG-Fip?bDuKA;9vG38 R,te_w$E*ԡ>l$|m|@[.ioPh92Lakoaѝ+i4[ Qf됺Yڕ`ww 2-XwK~} ;W86ewzR-s~IT=|d& o:\hŠ1@kѼ2`&s(9ChM0r"*4=gƴ9 H gT7ӾZ`/X־ hMN.=U}hB,!vHi##38TohJ&h" (~ 5YLMd1kIu3b ; +kA\"5˓}%4}z5y"qb〇 K0 !nH$)" b1NbDi!|~{:ofxHIx}QZ.,7]uf.*$bg!-@245I;Eޛlbw*/Y_c\Ϙjt_۟wyp>MmV_Zc.]no&={AU6$!_Ten&&ev|/AT}?%CJѬ&EM<&TQޙMߵ8&$+Q^W FapslEa- {.O(aDѦt=GʈwlFS:=1y}u8, G/$,Vtՙa r^sAs2waE(3wo~`fLIfPT$gy! rEhnjL5F;x!/g\QJ 5Ղ Z@yl(gIP9Λs_u&seEّ߹v"!k5T5k閜f-:8f֥am@\/4gW8%Tv-I'W.Ow<֐LRrOp),[je4كSa;8L!#<II,c5r|[ۯhkrO++wߵqh=$FNHstnȄcOa0$5=0^؆7$UVzpDX"v1)HDZHxIӸs'muf' c2 !Fi5R6&DZ*q9s,ucHXz.B;D,K ,yKQ0U+؝;ldRRK uA⥕>i Ⲕ](B~E< ׄ!&ǖ ՠ'`U޿`Q{FayH%p^ݏNƸG8 덵zu,N@uKPF0b>wLЖ0{70A@cC9ocq?MxzU.w>P^Js+=ҳitq;,@}Dž;z4VJ5N! K6؆4#ה\5ˤQ%~= QyCX ~Ql`}X5e]AEr:GpAc85_Ǎʽ*Ϊ;r*wh1G +80&L@#PGIHFShd -5oS|.ߗ-mֶbc!C-ĭ>ænAEF]>ʕRú1X,-%`05͸Y ;eEqtйc3;.;FW\b*`نX T HF si(1[SZ@cD"SAKwiwqI vi:0 IE Ƃ(AĄ(Q0Rqp̎My|G%\bq Fq!eBIQxr%2֕`&09ccpq*y9'xrq$9XP"*h2k[NlP֦%1X4"JgL:qc_W bsFx dhҿBf^ ;TL @֝OifxP%>g~-ԞEVvzYM$=J+.?ټ*;Oߐ܀e#կ~x#DF;?NݯAra3V*O)HţOwrY3_s RD0h1w=x'htO6<4&0lua6$NFO>̗@hMom~3beVF}d/Dk]6I@,f}d>U΅")xh!S"5{2WcDN )ۡ=^5zǬto1Zdo{h [s!oƃf7^8N'qL`6DO:FxxrUdgkKٍ +σRu[% C] ;zsͶF'$]d :ܿgW\-#S8[%1o[#ѮU6He>5vp&h#Cj |.g %U!f0&6E.+NIINs[lqOQxJXXc˦(8]Xy]_ ) L1{KY)e#@3\z9)2H"AGXG|.Q8RBH1{zDpmݢG$GI9Xp^AquiAq%d.S n<8YKP[IdZW F!)xU)BAͤ;]5hEUpB,EHi5WK8ȳe*ׇK쨓Lq6u,YF+ܝinn"֠k~% aݘf\w2u5pJ_^\xfߦ_QT 缗TeK5bAgbPW&c.9KwnSʪJ2׹#6l‘ؑ2[cm//nKK=Xm/ n+bVf (ǘ";5Fw%ov}L#j6+#kO^ץQWbmѨSNX"qbu7b,1 :SipYϥm)>&V--;D}{RU'@'8e= H"whO@KJhuILZU5&H~rM%=3{ÁģccuJ;| nA0n0Nx늱Pm)xţ;GxJӈRZUciՖ`gy@wkG9`6sSYOZsŁ93 WV@Ç:gBzNNoZe'G,J@ېNODZcQ|\sS'n"\u ʍ$z}r'mӲN_4On<\ zNޢ1kKuVKN33Q=]L^^}tWj t7)-{aog-U\La.TLzgnr5\gn(Y$j*2qov]vMta`+Z~ТNثvv݃y:ɭf߾{ 5X.'3[hZx޵B` =^'88Y&/ce9?CJNϐKwYp a*63.gkE^=o2^aw}x+٢$(w̦E3;&C3w妵Z]dH\vP?D*Ȕ/u4҇RL~lk%[BDa ,"LYF+U?^"'u=ƙNed^73H.Z L#Q ?m5"o0j5iGl_ch@!COb<3aq -bSie=pvq:%o'Y%!9%C:İ=rʪBPB(-a}-򈝸u]UX!@ɶշ4ݡwYɶ_9!C#yw7 l;ʿ|I?fg|HxI-- r.g>! o[,?@R@ m]P,9H0}ET\U_>4$?J6ᑨ-_FE@Gz_ G f\g FwnnN~P\= I9l꓿ X(!27,} 4 "r;I\KXqB[ *όLm.1k`{/-T.Y U/Y"rB&5PpR Yߛ#2ҷt(Q|26B?̂UA~} cxEWL|ULW]>pآD!' >yp{G ()&bLLwܓ }1@lӾu{l"X^Z]G.Ǵ(rBRj֣vܓ#"@xےa)Zus=( 4ZqQKʹ 㻅0[en\7[UvzW|Auwa=?Wa,ƕ˚id8$*OxwK)fHZ|o]BʷAO|+4:e^p=:MĂA29VKVf)N[ ?Y i+ٯ1njQ>Ys@|O&2r3Ahp ce,ָAxxl:a~R lvXȐ0XiФY^Mů4fyT@2v<I<լjQ/@[8›؊S#+`z,p.4,lw,Az%+&^.n4ﻻgW=<|-}? wCM•MbL¶a{}u(K>i%/#lmXn^9~FB;:,c&L-ɾN\74]bx̄}b*,CP!5VH'iPJ5VI֔.kVkZH˴Pkd]D:`=T @PʰxkZKVMڍFMYܕԯxl Z@DѸ]iY@8Jc&#8~0ɉK5 d8_և+<&!)֞:-(bj9ŘaìJiFP!(T=~q+7-{nv6…J+6U\X@Z`znXMC#IU&g;Coգ^FEe~<AZv.X*Fe N#PI@9^pt63qJ̓"@5=#5Аk|Q \Pa7]ɤQq0KNy"1|ACWX! 1ha%(^x㻄FHy]^O"DfTj0%!ד<V.]L?hm2Ɉs ʜ=*s6ٮ\$nÔ:uc^_&N|25[DWfjl F9"a (Bp(\n>.dzE}t3/b3r~Ø'ˆGzn2,Ѡ0;jGPqZVl 4*EPuc=}ѩ98e})H Dzp2!ʽSүVCBHɏ͇ZGX?7B!l7H_skF] (|9Ca)6 @[ v:F W0%4GYfzA$Йg( <FH>Ɉ$0}B*,9 ܡPcs$Fj )sT9ֺ>i{C )0Uv,-1n,f3eI^A FcG"\ 'jqO*8niDcĄw0GϢڥjk@"<֐ҪaX5n߆PPV%ilKyFy^35?(m,1љ=ci#4N8ykGeȓq4jO}YsjInawT;Co 9`uN2=8c׸NduRgV]omI~NN9-ivZH|FLOW8Ӫl+ᤋr?=%'DKЪ6GJ#s{qZ> IAsvבz뉏%l.~rJ&f;e.|nedRG=)## S[S'2v,2"\H.vQj/a;3`w nz0De .'7Y^=L !*(+uɈ.|^FB8i쥫zwkkh7V[Usr*aV )Q[@ԃG ZU"W khM|ZbJ$/}.&;o{ QX]eRl3@8N q6RQ0sZ45W`% OW;k㌣LA;d)J(cGit#S`!iͦ\IkYy]Ǐ`9ǣ@5!1_ C1@hd 嚢Qhߩ}Fy\j^/zQgo_m` WÑMւ#pu.60=Q0as6o/Ģ:e:\27J'άO`2ױ0<j>&v0ձR(ݠ1!+bǑN#'bIʼn s_̬)F;4D7NFN N^&k/ڜkoƛ \m'TkFlHoGC$HiQ 2u0V c膠w=AH*Hr,Bi4@ kOz^׎$|ݠPD݇TIy9'-1׬p`=@x `5Sɳr4-A;[O."9ѫA&5n =. bN`ƒV_/!;븤t6ѯyvɷ\dv=bS%K5RbK0䔜2>ԠO {qR .NDwin]*RujK^mEaO|L[HEbж+& q#2I(`|hGbu׻GmyU!G9@01ޅه*?ݗ륉7%b6nLrw'vЋWVsp y񛟧AD;I9]wzl:8oqQNzY֟|JfqGw9-";ۧ'?P%'5#]Mb$PKÌ )]-ݼUY#݈)c%k*R;2x'x$B Sq9JMҼ0EV]. C4RR|\?4V {ޕJĢ=JUp+u}* p))Et{Rݞf@+$?,X;I$$\!w^tO`cƿn4\"eL3%QcSy(2^0/Y0lfQja=p朁۠%wW .P 'ߦ(&C9DQf\o.ˋNR돟>F0EpM\[Ke(奲t"KÊ7eP;{ #=U s})k8{ni? ҷ=ӛ4WZI4W`tҷG9!:F9biҎqVw=6!Fe! z5♥8_7g2LJms W]\r$ )Ntq19r-siSX̌ /i YlLb*~Y~}47X݃w gzv򵑫ȌVZl\. L3=`p#҃$Ђ k)U%P%.>n9-/_`plXƴG%{e0׆VZ;sTgD𼠨5FL|R#x9!&/Ex@19VLuY \Dž[kqF ,A m]¡-F>\PNdGA(ԳG8=gC1H?bRȅVCyǏYIL:[AZf-lP`҂ CRZHxMqDCes%% Spe)o@P)6(\J3؆$ D? #5gNj;F*f`I ˒ AjFyAQAw,$*T09x5PA^i.XrѝPʙ8BEyFV#]l^!|r um.E=0uZ~燕_ci;:Nǻ /XUT26tL,ULc L=(p)%aH$hW7XϦl+)Lg0Yܮongz}Qlw)^\3[_}e1Yvy E R1a/ 4VX]/Y׻cGiBô/)=u#s|Li8vd! tU6_RwI{*uM'.9Rp;>݅I,G U'FI{J!FZ(G .Hi?bITlHӵoD5GsHӼ/t<ս'FI{JPOqbz Jtٯ=6C'=o3n@̋[e ^NHz^`=0M@{{t)ϼC)Qrt9-h Ԫ t}(Rf>]?\߬~|/j&m ?HK9pA\1,jcU6w\$bI$*[[pnݕ+kI]YMGԶ?}M wa{Ox=8kNl=Gyp\|+u* U 0+Ak~xP7aEwZV>$I+%0Mٌpiʴ?w!}s=;2ip4GLczŕޗ`31} ]c]\a: J T[)O[+Ά;~!\fk0ß됒ڮSOj4 >zhT[?N3%[Kv3zpO0c,YP %st3zpSzQ8htV׼ >J~bAHqᜣK*!'Tưۂ)"ep :#V s2j-!@X/As(G)L[|p:(ND3ZHl5E,CϥmcR ʓ 8څ7)M@x\jeJW2=qK} \ރ%q (" Ԃ6`R1tYbe$H %wGPDyUvut7\#dm ^,gk. sUo!R,1e^zILuɭ'~=f^ί; #p%tɒT KۋY,ujǒQ8dIZ݋>WIh; ԂڱWɮ^SjGX'm8)dKZ:œE293ƾSˢƾKPYQe,!ttҥ8y ҥgTuJtpt2,bɔn/SG']*/ɘnV GjL#JX3[ Q% /_꜡̎1'{ J4aQ+;aZ !oFLj)ܵr_@c\V΂;nCɫXg@}׫'>_KZXq8 ًp,>9hD?Eg=*ww)2VAĕ#FbνicSA2&|txt E@=,6;0?*Y>}8r 9=1J˕}=OQom},FO8\/޿FHVoțߟWƯ~^\8DDH#gϯ&tZ߭ng3c(!w̬?K}fqGw9-2+~==j|Qa$X]M=3^q%2I1.i%beFפo J@;ʐcA ʂ8AIRiV1So\0Mq`B1!pjQf6AZo~Lx@Ěmg]a%Sm VMٟY3ΈZ [<+DP5Ucn@dQWQ}Xrsv3pA>!Q &>TM+>QG"-YߛR,9g!"ֱvyBsRusL[iC {TPY) Qaq(e'9ԇiyI9`qp%<)//\=R%jU`ŬwF_%HfѸ93 c{I-In߯ؒ,YluK3Y[n_UzpC0毾]Ō78oQ #4|W-yC6R3ߦ3BP"z|Kx"&]^s 4*UE׿?324O|mn'HǍ?ۻۍ|qo7yvc1+x .›D1)N%m"gTd`E3S]}uF~=|ŬRUKQꃮ e+E`fQkvWKZBwp RGA_s2Dofznoד,c ;prNN#w4LQ'5z&,)KFx BL|Yd7R`KÎ M)N; 'ɚ=V<1PbeCJ\Q6l@5^lj- m WqjQA#\tNDiFN >k`J7?G8SbQb}[l kW"PL%NRr52J'0h.%QTG"EԠCvݛϋvhb|tTіX+bQ1r+HR$sAR)]+TҊ3xF,ծ1I6e1*nuFqJ_$_!(!Lv!ƣĉA'ܑ)ɠL0Ԉ` Fٌ\ٍiT`<>sUL1#h>qĞvcjܫujz%mۊ&CF<@#OQf{=TdhbhF :DaYrLNꌂisFGlqlhu$k*TThu$:z;9QUYmr "IzCնGprV4#AR榒B]p; ]$191szH1 8u!Q[)*=jT?pm Ed!5kgяZ(yt" St\}vl|6Jsu !^-Q+Đ h{$36x<ՙ _5xkjfIfbCXZ*I+&RudK ƠEE'ā=G¹C; J$l2%h&Zczc,_&)44gk\!-5PIrU1 zͨf*6Ɛk"[.hPZc5c Ft;B ؼ7ڰfe%*BI״c^j/Fo+fI{Pq$%&wJpz"<մS22qW.,dT]h^1' ͯSdJƄ5'AEDZwM X,<̜9?Hu1仛'݉nw2U=M6Nodz'yv?ʁlAuN%\E?)G(`'g2~b7iisg^`=j0Of7nO)-- I4l- 44ks:~e[V烰/ tT78:t[iT(d_GS=N櫯'Z g2*oIⴅ  H*{N+lr;U|g?/,!Z719.j\[tY/zuِ<\qkj2q&{v꾾J/ G .OxڳX6cvAKA0QC $_,8^)! K'럮'K#|nssdp'tV&_{8<]CqaǷd'1!akFޭl\#^͍,g<3=U z}gg6-֚3r2L=2.ߞldc<07PݮWW·gg1 OAQ,Mٺу3vb=2j#ƄvhxU$gS̶ڗb[2KgI(468&AW`ZJI.QQ^tSiI $%p# `ymr5ޯomaՆ*Vɲ q^n误WInd1PhGw:0 #?(.}<^gF+,[{lq{Cp<"-ڧ ,J;Λ.fcb7ddrד!!Uv쮅}g ?9[ǹy1_y&,xCF9~e!; YLܾ[jae}gqhϓ;Bsy+=IeJl]w:?~[:?p>㦅/5_Y64"Y ?dQy36ۛ:-W;9bӌEё{{ExaC5CġVNj|`C>|^  ͚a9!w#v]$Nk+;fUh\U?_TWhMo2!M򖞼#lטv"1./x|r9l3Ny`Yff]FhO.,G@U ˇ ;F"BkSyOX% @*jأd#oI'ky!U/Pn{.Z$U΁pE$M&$r۠ :)"5wj %Hηi[6-ަ@5޷鞶8#E,sCs Glq!w&ή+?T8wbFrm.l=OU>BsAɽjA{EtP'K٧}vݘEiK Nd|ߡttq%^͞킠;|v\&kc ٜr? wD"8ia~VkA<2onsQw5ju''K*\zV>]Tnti\. V7^\_*o~2Xtl؇O׀=,eǾW+?_#/E!UT[hЂ"j0]64 [ۯq*$䁚h'\;ov$5ɽL;DsŨxA4p煲ƄNe5d⩳x hv!&D?y=v(uDp))%٢w$Hq҄֕&:4"ap:濐!^-đѴZp4)rR`wڙ,W{\*&#U2q(4f5oK]WmEoڲB.BE܆J) ^VF^׭fu8\J[p˪)p,+=NKPsӦҽpخ%4zİBcb%qJ^[O]BD&`'3c ~rAQlYX:zUQEΗK*c+2.ttlʺ58;@ni>D FW~3}0IUJR׉8:SCyb'K:%Fknx-X.X0iT<*:TeEN Ф|:dx0:I,B|-i[d%}ǖT'|ύZԁ$WU=o@P7)H~Tzu[Fe\#34@XX$JP&QXAH.e9IFLݸ#B>ڋ/wA=^M-RK Fq66_vpJG%֮,KY* ~.ڡVMx PI$WPD91$"(M"MK*T%i r,+@CW{_]Yo9+,)R=bt4VrISU۳:Kʬ< RQx)RS͢\TsC;ͷff=${ӛ5IvKNnìYF_*_h[hB[-FAG8\\` 1"{5̠YG݃=ZﻸxtW:{{r(#ύ֐ѫQ@ ^å@[H<%N7Nx&0EJzC(+NO"(Y&@I\cR+IbҒZb]g>uCuFcZ{c#u;&40MJz V[K/CNUFhjV*\htݍVFk1^t9NnuqxtMr*+E^"0R듑:p-jtXk_:\"zЅEP|E1"ӹ>p4קPmTתk)%jbCI;l;fz\i 4>|n]"gr׺b7JkҺJ났K>kv3%E.wp_8_>b(<"quSdtzW>=֥}ΰknSUtTp O5Z S0S>hosG9p@$Uͧc=zhZnj{azݞ,nw zpkiМ D2nO(YkY~0rH2ߺud41b$YBNB`۾?2O FN4-Y+ c<|Rj*oVhFSñqyგ2[h8> k U8Qn*mUbْb,_wW- dBXV ф(NqӦf*V 7 KeWZ>ۢu%ke DUK]g{UvMd11MJLz&*Z(#NUҶ2QϦI_n=H(կ{]VZeJo?_`D@MUkDOA$\\,n>>|YXƄbI4)-hd0Jғ_x~^pIu9y?߮ ݗ6t9Y9DY`j g-WmNqyih"m(3lƿwu'nxQޮWyʫU^ʫtYcBK-(x<:xKVF)pJzS&E1V 5妍ɇ(悡=< Y]VS\Iɴ` M%@'̛],xUB< 5Zᄑ(:$"IDjv\DKRO%հd9~7qF,$\'߳VJ@X0*@ &!I ɞ4?\x&oL (L:+:.O O,'eI #d`gEE@d4Y{"N5' Mm_}w#Cfrnsp:5GrXæ$ܯwh*4?mmE#-g\ Afap~~FC~{ Š&o/R~A;\h\p@W'Ay8dffrc?m.q1l:!frAVu:CgX~-vb!g^?Vs:9[;_r} SSROokkHF5ϡ5@cc҇Do :Ȇ@H,S2T5Z9:.Ft.sFB(:5DŽN:#t,0*9" Rn:A*\ؾUܪbIBwkˬxCsYqq4n_6o>|'َJ&&9GB;Xb}WSG+5OD*/ Q;wB@% 5w@ ReF Z*|oŋy|F}RbHL ESO9ẸILvM_|np::.@T X1Cd =ub>R" hiґ[S~;9M#D3b+Tf0 #$.i5Zy*\b@ ]K;|6Iu4ڢ9I `NPd*mB0D΂Ka^UYPnb}The"x@] jzL& F?ޒ(LJswAR*[qrg,0$g2ͅra唨\U y _-M14) o>P%穔g٣'n齋al>?.em!M+rM5)]  } n=?$@EC_Dc]=¿hꋘ]׊-TRzuQ0b>Q 9!b^fdW xF1#I 'MJM9Q$^vΦD+ Nq/8̧!.i޽iW6E-V|ŦSR JnŲXsJcj+3wv-|z,ne!>FmGePf:x~?+9NcV~lZ ~{ :kg=99 ؕƌ7\ζ~6].Z齺G s-*e1êX>fճUDc̋ٗҷ>b%ѭߥm. cvʺ7E~mߤ\3YAγ5s?-zJ͎ FG r9:3@:S ݩWex5Fvᦳv=VU` [31ìdw)[;;0Zkr"ٽ&D;mJ!?/|wѯ9*mxDFZlNn?LgN LUyfod#T}krJ{RӶS=m@1wx4y#E{H 1ƨ37Okma!4 LӼ=`N)? G:2*';',靽RۃF D̬R0SJIXujyU.6^? y@GFv$mz'ARg-\h85> mʌ]라 GOrxl<\yq ٜ/&yʏ=ǟ-BxUj -SQ0PҢ+)T7N!'S(;ڮEO;դ׋]mGtSi @];C;ϵN1S˗@._(NǶw tD^c:0RmzIvKn0g"Ƣm i4Ř|ykIxTZf|:6NԔ+//1?~jTӛy/9.\]}K\M/5:ڛ\NLn[2xj?(CE~luM 5euܟVm>]|f Aּ=EcVAJu0%q +8]hbi )GktqqLly5gs*PvGoZ-Χ.sIh !WzD%WSF;Nb'P,]RjUy3#]nO*rK>wiM^MMim*s͑qtvJ̃"y1_V;ae}v%폏cʚ6_AiR3z_\Ň2/ыź4pA-b @q;KO'%g+{I~Dԯ-D3H Ͻ=ҰvtS>҄1~<.T"u:#{WCxW=9G^GV_Kh4K|TոuU}$@dz{IBG+ {IomZ{,TἃEΪz:NNn"NZ=Hrq(Yb:dK#llD(Klgoc#MG"G=O`nCsPhN9,8:}O9( L=HGQԄCWC C}?䁥8Z)@w"dLsR%+ rۿ/zԁ/&$s1| p7Crvnsgz`98TIL|SK<-fSň3O ۂ4Ƿo20 *3h tM%Xˑ :O cёؼ+a"h,hm~Q Ds*96 \pP^рom`3.p`8%LK-<0a%!h l9'V ItP.o &@QM+G1# 42 %s5 S>*r G-qD}Dj o|~LZ̥blV&-漴d ?{_Ҝ34osܧx@ǘPq{TfEouaޭVޠ` fSGD@=ܻۢWsV疷MK)ݻuK';/s9@FKX Ssfbne |E.PkT*q5.cKp!bI6AZId1GD#\z qsc .hmnuFDNA~pb%kk2\NphvDtvDjvtvrv$dv$}9w;Pe1nrŲtb)pbkwwlwwlwwkۨRsŊ,"&``0iTwby.TD> Tk'$D"~GŪjŀ㚍:٠o z=/ma,{~wgo,nW˖mt\ɏ-&e|{l86*>H}~qBol_LRQg;us y{­KoKٯ8&T) R}GMR[*UX'u[w&TVӵҳn[2pM⩦}GM[*UX'u[wHS[2pM).^|Lv㚾햊AI}G]h*lʙڭ 8&&/>vPO_gAI}G]>&Tiڸڭ 8kr.mD *-euJ|<^ZvkC΢I<մ鈵했A%I|G-Y}Gimiڭ8 =n햊AI}GaT"$}KE ESxJ#k i7EԻbPuQKuW/* %*QZhOqkk7)KnTawn֕âgBhOQշ ʉ&c]X^۞[!bk۔FeèE%cX/[ݳ{ F}DIk+P[)%o}``buxc^ [ !5|kڔM[c}XӪ$(zHc0kAP lN<&+\StB%|/盋8ID"—8 ܫXn.a;[W^.Xt("Nz6U): H0A/FYStONWbnP?^ڷe߾W^HÂ94c+*ʵ)[zߜ|$P*AA4րG#?l'Zf80Oxu?@@[]-!t8jYPx i~..@HEF!2*(2eXCB ()1HD*RҘ*  m)$D#r5~!!?aXlw cK89cT_<50x,#XQKP${^W@&kof\I@SK5pKcEoiG=ގ\|?uX`v K|6|"È~!ב3[|lBudT ysvtxk/~Q}M"b.t#IZi[, | W= D9޳r >1H 4X񻻳[ 5>3oa|Lpui.aKanϭA" Z:$oWdK1{vKY~=7G\bNE^=-ҵ%'vm>%2Z0=Qgs*N?{; 3j9H`Ww!ͧ*4֗_/cTP{q'blxjmg6mr^:=D RR oۋkusȶ2E]z([p 4_ϳe*<͡ŷѺMnx]{.Ԋr*=ą\M~%QL!B `#|`be& M-jd-nqbY|n@ S ݷ!Ot>` l \KvFu{/\#G9q VrW{ SxT0 8p# z;!{ZM]?"Qpm$쁗uCF)xb}12{O|?)Z^2̱g 8}QVl V]=pzniXpP66y%|pUĨ ^-Y}&!_Ga!c jB"g`ܗ8.ArbHH"ʆۏ,Na9H0luZ`mB@C렙AsĤ_T*:^wA/d .?@vk \ .S2ͩmWSYOkz'j\ 弴E  *ۀ>if2.EINQ[-kmH /n~?{ u% גTl'~CR ):$Q Ҝиy`| &Y2~xL _~ï}y_$#y~<=pzj ?Nvn:m@xz:vyOYs.k-䎣.9ϻ_W'ߐj]=9 Xs@/|~(w!@Td| ,A,x #cfjXkaì Q/oG67Ɔ+Xt.WZhe`B зO81ǗJb7KjM"#g4UP,llXm[Ц1Xgu t1};{0>cM`3C$AW§:lp&x bO&JLx*."%ـo"'!#sD0̥DK<qe z*G26)dqc$ܱӔQFG{p03`15b uY ;OQZ #4fL&ҡRU;Z,ce'Յ,s)lD~ mA L{81xh}6NL|8 3Uyr8:}|;\`>>y8Il"v] ,;=V ~R{3܀9B(pnb`;U2ğ^J F鏾>$Pl5Z\o;NyR8V]KIؿƮ)3{p-jʂOOc3=La³" .Td^i%.O?a3zQKwf.nS  %L 8X41\4 4EF{y$Uc5atm NIJ35]5BED]{N?A:~NGo|:)W"@x{?\I@:&iRFuwP]!Y׻7MxӥCİFx7?I9a# Fm ! 尉34.C.䆬9o [wKv2A{p7t,gacU1J|{|<5Uث##Y6׻'ęJiv8z}{48xo-c"(x5W<^_%y5ſW "I? Ƭ4@iB{jF_\~ K4(]y!:ގ`eoAANf{vo:mfnn2PQh/"==Wc…TᨣyQ/^Vq9JE+˒oNt90֊MŢ &?DƓ,Bj4Jz{Ԍ2r(4',nU&{=!R)wT"/Ac!Ā=Y"Hv0Dl%ҟH ǂij.Gc ""\N1WbAmס-+8|aI1Z6t܋lDQ!]ǜ(hs*L4ٷb o[z^1:āj-%vZecGB緛(,vq@8iWO-S G'3RcK:Wm_ gAv}9ПyhPZx-9$V%FI/G֢}HϢdqo n`CXHB5(ygqA4D 0REwbܾ ^cVфJbafI~[-JfC#nQ*B2"CA)qp#r\{~;z JmD`jz$+Z.@X*6GcI.Jn!Vr]2Za1$XyqWF{$]{pt2} "j`Jhh^y&\RⰲyǠ{ 5:9Ç g_)uD'}8{NA4?cf GLD ׂj+5ي ~g7DRs%Q{8k4sD{ޔ1Wև} R 2Z( rBc^vV.rۚ|$L/ʨp n^X8T'CE-&Ӽ͑Ϝ@8,F!iB"#2z&ioBDHxc=<2-ݴ "܃&O[- vRtNX?]QR5y.ObYejAs.^/X)snގl/Dax;"Q86~=<ξk>cFwr!{8| OapqS~5gIf^1AC0%8YvV0fKK5#L!=2" L4T1OIuC5NVԎ;vS \p֫VA0IoD}Ng9ܮ9JI9<8˻Is8Ipr407 JEJnNJhUB<k Y*d;D ՘j7 wa6iH3$/vNr1Ɲ;]FUyXnpƤ,րt `owT"YA~dD`(Fk8s F'`Rbuْ LjJ{q6Q4bbyGW?wlC; 8UڔN~N|~DP0- `TBI MdXEN+v̱ڡ*ٲQ 6N/ֺa(c*`oGq͜r­7)i.|,hiy4偵:gI{:N%Ӕy* }r_B oi vlvt]qG$/ġa㶿Uo[7ے)$M OkT\KQbUh%AR2^ps{Z˘5$L㴘ݾlx:+R'Wa  Q j+#CU9 ,C#Vޢ2M u|;˔<#q_;xՁX)}Jѽk8Jvi"h-Wr e-:1 oYjBf ;ه+:w!ϪY)$[O Ka(Dp>:ur=#vzkC#0{8,>ߘ}E<`6?%Dqm;n:2縬?0k)Ə>f 6Ib\.?5dZI,)i!-40z i }Q.٪V_3@Y{AY<9&wY 5˟YH9_/L&P %LOw|?~?8Bw/d<l=a-"z"ʬ~Y5g>A]ޝc_Ŭ}P\J wRTZRFi ?MCPqn)?zEʙ>WRQ; qЈV F+Uc[Pj^+2Q?~ЊUZL l$U~J_RK4(4ivӷ}6E.u{I]'w]{d4.ǚ(pD"+gD*P!1&VKkM]JAӉ.˲FXrAA(gbV9݄ʤ`O~d ƎDiiƖ#\ZX 9)! Ձ>ά:#fM.p-q.?w߿&)D ]53sHv 4zج~#ֺM-Vm9@B9!Ggd'||gjY(aB0tyc]zq^vWjb!"o9@tS{"D{w;jYQzi/׵jf?aV-c״0EuC8IuΆAqDnu^% ^H'Vd?-kt\שFuq]Q~"A0>/p"+6ֆD Ql{o2;DN~tMrjm{v-QHGJ_|HD~lp JГWyFqw~R$/fːPUOC>`<}vynCbFyK-(=)½АvpCa.\.|u;<9ptMGT,7 ۑX!'r4a_oof?w9z4Pb8b?ZAmGz._ نqq5{ge0  {J.i $';{9ȡ#a|w譟XU[Va2vWOF=_h9߯{|8VF{\d|It*QR\٬I+B Lș )EZp2aN$F3[پ|mbM>/؛Bt|].H.3^Q!zLƀ1@IF$!'e*8ք~AZOҠ]򼊧ߗ<:WŒ#eu!|Z9򜴁 9i4xyU ȯ>xبoG^p3Z[QM׮Hg@٨ DQSii#Db9 !h4D J3.$S,/B9Pxθ jyRFD+ѧ̇Mt4EՉ*1DӈkE Gcнbj&FXpHGѸ`i(y,5*҂P4@Vc #%5s Vk"h'zQF6[,^euEz%v;H@s3bOq&|83ȓ(%b0uJ=Jlk2Bd @2uT!V{"qwREI(|EHs֠Q5_TLD\ a֒^?O$ zo)4*>h\6MKZcų_.Mf+?hugibG|q:h,j %\3fD-Q~h p?pU f-D؉ϛ 1! X, 2ϐfbf"-ں1 Ҩ<0 W!&]ݽlO9p™+9Yfdbս}!} E]4nH{{D~U?ڵ{2:[&4Fj"R_ jP23e&&p.ܴ2'D'01UE2jq@\~\.(66A鹉46t&: QKPȜC#Vm[§Zqo;{M_f6KLӵV*R!%@:op7>ٸ"bz}{n~HgaW$Ra"Y_am%N)N{ɟ>W%wdւ)>%]tU[Wu W?iOU|˪ːԥz../b\Lw7-:4T;pM[k^CGcp:qэS%`΀\<J 3>;d6E,YyǓwDNixV/|` Ё6?MSٰK>H諉/5$麳<NzZui )A(={wN&_iu/)0*ɾLHb\ Zn`cPDZ؈B:k#X((s FSXO9Id@pRͨ2(h ~#BrT!PM٥"2PDhPƂ BFE@B+Q3%+SPå6NeڲKM!]6rmfe!|Mkl4W0һM)nQ1^@aᅱQ5 AE lf* g|*FoB1[z/~\{S!ȋwG5nL2)IMK" e~1j>6-Onzg*eǒRga4_/\;~W\L /I]))@F 7z=~f:'Y:iiy+5Td|%o>x9X&Q@d',2YQg%%$eo6a] .쳐!5;X[ϑ>9CT6syoՎo65X]{{c_.K7. /7R ׽Nk2WA( =8kUS9~s|dm53LW_Ǫ9Ywpnۃ1ȸdž" {">jZ5Fzx_c]CJQ?3):.{Xr$4 7% .Fx Lx=^U\hw@`7+v;~lb'.` %5}K}Єn<|2CAŗdģcA RYM'dSӃ <u0[]MaQtީoЀXk3' ] hʩLbLg Dgp }B+.L3aF$7_{#bԌcǔ<);%w0e#+ CivqzZ~ _pFy4\xew>:Wxo)390(T;R#jz2c(c$jHf#6EdmFuO7QDu[ju;А\E))PvXBt\9iREB]EgdJd2/&zS]IFQ} fpϣBǵ(PěJ/< Ww@wsu5^DZNAOlJțP3I( Ai #8/;,Wf([k̷Me"&P98,w_6J>e\ セj=7o`35o_e d=/S #sьBh9\ #;k"'R/#L^_.O8B' OGcxkpz\kdy= هLaDjO. i5G5_G9ܵa+oifXʉQF;-lrvڍpsoʞUCWh,1Iɶ_b@]6h[V(㥀L=$|P'%t׀o:m8ԪME qD*e].s怬ԆkdaX?uGp2Z6@dwYle+\O,Ц@;?;4mAuXQtc\:x[G.c-o ^㊞^Hf zod_}bDcw;?;pA%ıVn=P{pt •*{X?IWᤋWV浟, }S`ߡ );W''50,JCHO|Znd4&i=Ͱ^eT ӦiU.z^qCӼVCLϓa UQzӊXdnspuoLT:}+ ZZ1K`}GJ3H(u(x_+T(Ɓ9?@Pkh1x;Э9oNfQeI2sm Y2v|{b2T\ɔ%BuF'yHZFUH\G=f)Ij"4J|r7"JW!+$ rjJsd^¢0h._zys/!/v)/PL^ZG(W eR&QI cA&@`n_k-cT !؄5*p%pXoɅ ̋]%DTEY2P7,85cA6a QtCHgZltQ;)83SM)V3ke'V_֝W/zʍт~VGj40 ($/ѷ#8ĂZ\HRS艆Nz=i~Rӄ!萍;֥q6o]]堟\ڎGқ(OC?toГҏ߿b^>|?Bhj^aO_ruқTQ1a㪿W(^拕ʽ}W7H DztzNk;_Q2C9b#b6=~?`.8:P"}9𩀯 9~S^gƆ8հt[\w-Z=8H~d 472[ Kȧ̘>#)rX9wd $yz]9j^+Rف)D,d0\ߏ;凹Z߮pzefЊc8K1Qh8nUJOrJ ,Us%k{ɀaI(gإ:R)L',r^kd6 k)]Wwx:gwGtv!ʔF%|0^3m}ej~fWY^ɼFx%_O(Y<_n#櫕89yr5JA4.R90RS@Pfˌ?Gwu$ xn6]w˜kS.}댽l]2p ΓQ #a]L6j}=IQ\6NEf\Mv}J^sŀ9#53*O/ttѽcߥZNQJL?:e#+Dֶ1$ \~]1z`9-BZI+ϭDb+';6Bx%)x "yLnαesc;%m/ FgwDA<4-ٽ{}iLvP1av75Ӹ>ݛ+ִoYggd5pfl]I+w(C6ɮ{#Xnk0N4aOyžKo¼?ֿݕSEÞw[B!!{@?nP{hWQyOx7(T8]`snоֳi)?$].Nّ|IgHLCW0vc-C;V "%ndV4c|sy۵0\smLó^0†4B̎ *씻7h!*or\8%~A㮆 HA H1zHKJ<dIQA=}JydW@3/c 6cDs/⼿EW%ALx Mڀw[-"dPF:;)]_ ij+4}w?~_,q Arw#YRAF.#>H4 с'tGuMȍΫHbP2zG19SSj*I(Ro:bf-|p+2ѫs(hS'g/3H/ aVTDh/8TRBōQ{ Jp*FNHT삙FuQ_kҖZ|zO"}{놳Aw$6h/v!\2cm]د&P9(O^;;ђʻ%!1Be ʗqiVV_ "YڧI'՞w,ul%kA<{Xőkt JA s^FBuOSɱZ{k@Td~ UgZ!@n_[/Q@7qzФh@~I0'$9a h ?E cͩ֬ښusDܖTM(GDSƔ d==F*fM-sȐ(`ʉe3RQI஛o޾)Cc5@s^NT4'1xSB1&(IBerMoHyVz;Y㣮pn,Qf;.7ülIpŨxDקv?NO@/SkCnܣq E;{{t>ʷl>F߈ ~zW&Ei։hd*i-8}P= Ga21M>੟2Ah<"fW?뫉{xo^zWOxTMBBMF46D:/D&ͭ qQ R4 srug67),J &Qx4(h8A=Ny}Nh{d4 ìZ1ܹf bO;la2)6oGRPVPdJs$H|F-w#f`ni);p1i Z)ERy pݏ?Ļ\NQ?)Ւ 7_s)v`N؀)?v@iM{ϝY'<%`G̨0U՗˙2 D&7I֔'^5~lae*tUr4w,#ƲYe $Sa(*D3͑rktif^ t^&˨z J4U( 9Y& d1˼&‹\on-*n 2^tc>kUGg}y}G▁2귝po E|ޤ`T^qrRO?b+^<b,ꩾagL|׭px- v$"`.V|D#eD1%5##. ;uba@$5 EZ$MHhXՖ(xB6FE+"TVvARhi-@QJ'ż=Bݩ4e~ .e頔\ cn(giKmQ۪ãm4^QAr#ͨh-xzYpI1\pm4J=1.9c8@|VԴ`>yeM:&,f>5i$b{8!(XC(J]̧5Z3c%0092HA 鶧da(\-`<rbqQӋ0ez(LåEadTŊ <$5^樛]QH #^ VgD)t;rrnm&t6倶 /n~Lߎv0 2`ڷ<!_2 S-~TۢZ5ʙ]\R5-xcy[_41~FY⣲f7wxg[ϋPչSg ~zu\fwwAwʸ]M?$w?~> }D]-0Buqcs})3jG<ʊyw=> 07?_-\C,z8O)*|O-H+kӲQ7 R)2 ˔i)aOaV5f*r'I/3`ڷ-F+hP(8΂袜eW74h贿p8o/A*MaVops}J}%1>xS+t<2D,)iͽIg"\0DU27+@"3x/Aj|pR ^\SwxBWrgCL岗*|q5a Pm`2 kWzw=:ī$7^ H/$!P [si֕ #`{̈3 ύ+^*bzG70;?yO+J+eJ9l!UxsFq^[Qш=4J}dw攂f Sj@ȔU0穆wnm۠_ zD]JLRp,4$UpRgBm<1WÙ:Фj8S2 jt~6_E] d?_Y'O# 6SُdWTr?y 'ߖ7ߖ}|3|~| ^,XGq?v+Eq& _Y6\[G6~m>GM 3A=Djg1*ؽtFRoL~+~\>df>f3pJ\ro!op!CT~&Ad u9b謡&l%&͡FSu\i|wn*O3+7YSMQ Z&jǬaRZ32EAR#*Jmjd ㈮޾\j;[j> Wk嘧HŲF9AL2 vwrQ@1WYjS% K) ~iZ+$Wwח9x4%s;Sț<[BegUʐ%)O~ʁ/;rx[Ӹ%?ݝ>$$oW`(.c)4T`BK~m[ \9_Qo8`u?qU&*EjH ~ l45` W kp*B+~ Wqn6HΪ5~jKlݾ7n7|M{~ŋ,Jz8cI^Q|$Q& ߤ x^{|']281.ǫR)XcS&l1=o]N>?&ݼ& 8% c;OB3dGHAXd $[x.ܭ%3ۢvw[lBk9XE)z8j1eEm^q'׾^03nA)1!rU 2"њ7Q&^T Tq;XݫhKx4WP{"+':,0F'ԧJ*W97>udҠz94/e2$Y{cDmzBDojc8yn,H(<;Qq~cn5r?tK)[bq7]ډ,bxAFGb c0~2ߚyV faқ6[ ԪOMy+[oyt҈LI,_S3S:"]!D#.eG7>}g Ho*,Jci(x|lr٪;/2u xUƥNQT.hjQ*ٞRkl 9.Rx(_ @tU1 詥d:F½Z2QlXx@_e9:'ҵ$ҁ'Nd)pzgc[xxd]sDdr #w)a͠ b7=A. qcQ%82ƚq%p뷁WoiZp-m^Bya6?{^n+CVfj}uZVTG3%"SlB_筳\̺O\\͖^¢/Oޔ5gu1\0Qּ} 4OA1(ӓ,]wNG^‰i߂X(W:fTWԾE_M1}~}UC+-ɝrqAh8 4[ίk뼏jVFUܩT^ :EsB BQte @edX Sg^"2ƋowP0ݖLr~q5 pXD ctڑ-c)lf෡ב۬!C]YAkq<܄h|D^='\kVn<8#TPg u>a*f*@`e&[Hjly4;뙝Ipw@n l$IGKt^ d^GȖ1q-.~ZFj[?~8*ݚJkSN~=x|9(S=t@檠F:8#0J[:Z-I ~Bw^gSJR_ڻَh1<[\)Miǀ33\"ld S0o~3g*.C j)inM:'h|}4^msAvP`q ʰrswWHN]w2 `mm~Wz8szqu62h lwlap+}|cm0kzo vXAk*=ؾCp([4*ˤ# Af8r+UH1U8K W홛4tJ[lbx,H\#ҭpȬ&S]}*SeHdHc5A9 ކ8=lL ;< FeHu\RZQjP;EDA$t 9u:&gۧ1zDfчd^}ɯk:9]s*XJ{0i'QŮ6iqDlԠE`{e5lWMwaݭA@c IjV{޶% ̤C,kωvfI Ř$_1߷l%'8EzҴ앦-n Yϻ|xG U2qz$pvn^pKd޼3]qVFj8v@ΤbBR*y,Z 61j'e?YQĄ}t̨zTNGA*^O.(=3#`㧼 tRa8!;b*Ƶ5oztZјӠf3L3EY FڻH*=֋lOSDW4{;EXJBr8ZAR oq0q1!sTÀR/gҘ:JUPWsxPpei01u 7!0KWcm6|IpT#syGhVƟp[5)9a|Uh]mO0TUM^ү$Y x9{͘s};٬)B@"1 c*'P Ba 0d t$"P3$V0KC2,E^^t/^'~u&^x׋i J|f.G2?w[وc1\kY7mq~fsHԳJN$A # #0,ɩb hpD9Lc s )%u< U!F:sZS"SϔE^E1ae 4 @F67]zJmx3\6#+S[ EV˾j 倓ze}}}Xz{%+ip[۫a VG睯n1ﱠ@ 8 @a1cM1e+Šfx䑽. ުa~0pT$뢘2̖PWi?Z|VS}Yy!}}eOB2.Li)8A.*'ҳ g5N AO"7'Qɐ_BZ-OBo< 'КyK{KdR ǜKR3qJ<॒u2^Ay+9{q +Z\9Ժ ,&Wu邠bC祉Z@c"SީF^Gb.y-b lqT $9G}2O[ k_l!_Bl:?0D7r7l{52g.\}Mp6CvɑGM_Yp= qaX³K50 qeT!D &2aΜ&1#.^:j6Gft2lNqi4oF9(eF+_\YLnv+ı$N&; jssQy$RȪYw{TRPwXn7#paL4z5^K{3VDd# FG/$F&N G ֡`*66΀>s |%۟5̷ޯ_~-["0/1j枅Bc_ f;޿$;8y ]1Y'`w۰_4O>e0l[0'pgn! 'xĭesa<+,ݏ1nYyys5#qI)*:m@F:@feX 6'oz|!N3.}+faH9gICu8#=%}>9s\8G@85Haax )q%G^ԈF$ :&HZ [ͽLS~N6<0Ij9e]?#K eZI7o!tʐJqcIh *DqLJ#6Z̘q Ŭ *&KG9\T`Ŗ:8J3ÎO Z9Ü<ʨ RVFU K9DVϑ1 d¤4RMn-dt0gL-xb?*e͈x@0EYxs=e2W)̲}2u”y Csr%"/@]s$D\ׅ!aL}2n*5-](UhS3íH!YrN*{ƹ* }j{gpضQ)m3u[!#UJ[" tLQĂF*DSC:PJgW$ ɠ8  |qZڎdZSkBMVs9Fsڞ fX&LC0LeF18MYpm‚06Ɗ4E(0,RlUK6|a1ZTS׌Œzy0~r0P3Gq T2U!`,>r;.$3]Ϟb8uD*+_V'#?-#2jnjQdO !jI 5!ix8B5JĀ[c8J=KhfsG5iRC͝nsWb%JJ6u3xT299:tϪ q'p4Vc<"4%;0& D©C=Xsx:p5ewo2jB͸.+rH6"h3Oq[k4F:U RCRQ4LXo CdIN9AX ӆҾbU@|Ūs4 aM7;'1vNbc2VjYE$A M Oö$mѲ$mI1l+C᪁N +I UKb7 <-MbnpSpKbjkd^d#&yiIMKnZwӒnZVD^± H݊C 81.!xԇ@_J@րL/D:T qDRG"N9- Vƹ y yϋ+nv[!sj+s!!KIC aP@U Djq + Դ/, LCNqAOOe\0$]FBL 2ĂS[JV 쭐\ p>ZVo 8!e`1;/K.1z>L+cn#pL֘6)*BC* KR-gSGx":,qSS ~! Ӳ!UU@BHBgsfڞSc{vlO.,J!8G@Jb6!"M©dz!Xit`Xi tTtsYHIvl؎۱Yb,֝sB#BI `0#^‡HJ΅TGǂ:[wUۺtO)kf)mm3VI| ^WL 7x N^zo K@9X%<&s*-UU`Z6vhIq7 ˭]ALvQ=2]>ȤWLQ%~Nrb.b5ҟհLr*{ QSlaX*u ^lu+[O7AAss7ŋW 7NJ+6&ڼqWSKE6Q^77V?r_޽ֻۭ7׭ͼ9Aix^cO{6֒A$rk0fnoŪ8~ȼab<AK+qD+>1<)paV{weg H*>szznmmMf=?m [[ko߇gRs< >>ln{k2OI6ȏ{^fltn6]IqyIj$[];ಎ9>فxpQרod]3ǡ|9~h,$QR݀d8i0 qfh$o(|,~ʟj0~{| Nf~{./O%IA,W'6Ol^|ZHZ/#p3b2̾NP-뗹ڈONZOrn=g=?ʴg]Ifч?lq'.|u"= ~ :pg`Eq)sъ)ܳ~}6M+FΑ+?lLir&6 ac {~˙exQTQ]7yCIGvCf(@x8q MHy =D)0 \pa.IAA#j$Q0ǣgbA${PWM@?y^ ]]YSǒ+1n\KfU"07fʏ jcZ,jMi8~pz3++;9<Jx_qVT_Zs ʆ'ϏNro޷\[wljSKBb9'X˷zrKtklRMnjdٷ݄7\?`p=dctqk4D]_Sk; /F48 L/{ηFK%1&jw6L}:3)<,!U tRrq-:F Qr)d Bk1@B%* mI:O" %4݇4}ШMl7P+3odrx| U0e2SD9bijF V+9*)P2s R |9S9S9^ l>b iLsۘBDM75n#H!GO>r[]x1p(SԂ!B8gPcCYEtX|hu&ƥy9֥ڋ*rd+s=خS TqmTc:cB~7/g8WӵtCJ?pRl]jW n{m,x|~Ud: #*STWg:hÛyNڃI5[# z# щZ f[$& A&(7WJs1~+s%`̕Jz) P;#ymX Iɚ55%kjJ,dR >Z |IpcD5@jCgz&!`]0 h[: 'Ṱ)`m`tѤQYGxdѾ{5`FU>|f'2Ud-J*8!2yi9,H> J?zp09:29zNW ֔999^|e]!r@|x175nj ,~*80iP wKl+80RᩤCka P edנU9SN/L~4x0Ze& jH CDx/ Fėfn4֦1XXw2#jD!{6`f1nV,Šubbu+fxEb:LUn@ƺ"7SFM  MȨ`'IJ>N?0OoOgccD9$ć2Thf .Eutu  j~gW͍Pqp>k @9F4.lCG.n ]._hJ^/4z:~Kv8znЮj`m؋Ϸ.^8lK>&`$[]"\ml*B;Wo{r~<o ([gʍ=Y7ַto6-y\߶$H۴(nlɷ--t{m۶%+ш 6ZTC;e$L.`qR0P&VȢ!3.EZFg/ Dn0'PBޑcFW); y/XbҒdaլNeD$)5Y[LHZW1(9n̜=6%El ,|_XG(N|y]J \ֲnTuTo?ҘݚGw.wrv$(ͱל߫:>%X \Vd"հ6YD8cFt3 ~P%vB6BSQX/R* +N5s#[.9_ő*lU*1:%vj"@߱Ƙ e?9ٽ1Hz1Y|}3{{XHLa!^2Nkw0c1=lW,LXҠN4SPt5S-,ƗR>m2Z D1̗1t-jc4Q!dPw{]P%kC~<;V罧Fpi Iԩʦ pBʹH %*HlӛrP1jފYo*o)H;x%d9|]C$Jk[ 'un|IɁn2*AW8I]RFҏJ/QHe 7yMEoǻw,V(J;M.{d:gmWYͶņbbl.E^Y,瘺&H,b+ YZl%(2F?{ ;WL+Fשi#uu= y1IFbHZeЮI!HR$αP7PNYW^+c𷠼ϫ6݋~ևF ;Op{K =|*FFTQ-\z ҎưmeQqQr~u5j5[r:Ⲳ'ȿӳ?q-ztY_=`ko7F$*iIB(S 1R1{ zJ_4!xT.YEkuHHi'&Ό5#k='d-nF֚fdYkF֠ ^F OXF>8w89D+Ln׺gLԹS;5bҫ5&YOSߺH=) ۅ Tb6@KN/4|m 4sq:Gy:= RB t=̛ Y)GNoBm@6N'uf|DݶLM6_ۥ5Ipz|'9V- w|C]Ѭb'(5\l4:rц2e,(UL-q65Z1ӟ×2 []Rl+WT"I6>VcJ16BCQFNCjFTɳ^3k͕ELBUGz|eіYvQs & ]^`X7V\* [oDI+婳+`j#`9הpKd_1.Rdhf4jI]!8*c6d]}jޫPQM"8>>WDMn%s&8ILZCJnwD= ({J郎$R5;Q(u۷~9Cэ1'A{$hnGIZ)',y=O~H'{6(㵪 21.|(d5$W[h/JM:}OH (l48k\jh5}/ypkH 9V%|JzPxm)C0 x${DJ\EҧwQH__UF{HI9R$_1ˁtZ@u۳4Ɏi{Ϊ#W^dJnOGgO) Op|LP瑒Hڼg=)Wկת]S`?Fxʺhp!/%kY6&& Wu\&br֢\Qr[ "[ܦ,N@GVޛ=6Z+j$5TzÄIk_ ȵnw4% 8v_zƈl+wƾ}'yp~ܒ 鿇7fnwbE{j,Ĭ{Y_6Pۆru {g7g7&M! >r7XLې\u6/)aZȝY%+W˷źdzrz5눙Qίun.Ie*2Ew]޸h>2O/~0⽡<$=4emHEt] E1qFc\e~7$[zw$FmRS>fYo[G%qCOjb_yzʒ!|vu[>K^~n~W3wt?{5Kb׳?_Xgrw](oJcYZ[҉O+CJ~JeOf!JM8gr!\)0-Y7|{wXeΦ0[(ii3X[cexwQ 4B$[}\t<띯\{UY:aVAQxlȳ-,L$|2 y1M?]Ūd!$~CFVg> ׷{/WwtxȺ 67E`$(9 !d "jjZRوR S.YےlۂFSJ&(Sby'Oqwn1$El7JO)?(c$6!kh9zA"LrKu6-& "DM [(1֪P-n pRMAtwG%if♲ޥsn{>NdA7%4+CK<6B^#зc|L8ZL"YU(rt9Um"QN=B(s t~iK0:(&4^ۮh%l,(; 56 IRA+d%B}m)mk@l1 rq3({P';~wU dTr\Ȕh>k1'cmeGq0J#]ىQH1NB~ӆ0_wMamD|~-7"&JUc'MUEvIVC[< fn2d*;8Cq{!Mx6N}n:DXQ;ipE")Q*{|H3O=̝O`y+un]9nŹ =.ֶVwl/'uݫt۳&%詴Ri #b< z@b$EyVǝhHI-q$UVz7$5dP\ YjHt۷ը)Y:RfH+~/ɂ䆬p̏sV3_)U% H𯀤$|>` Yk!mߚ^+$,G{x'$ܳ)$?7ާwXes 2$ sn0I)0ѿ0(nŶ=1H9E]ɕnd oL:| $TKl(%n ;_(D()P|_u2A%(W[4ZpЍ Z;\iPC@8@&%A$!S^." d"9y=17Z=k*+~"{Ԧj:=GTb":ܟPlf>hK=~9UݶO*u75QLOBPתY0;{q?It!D_1¦F6%єUYHv5Y'TB[!]Mul( GW!ZMeS+< ';Y<}.<Љ?4FMu| 1MDOl=Α2Mپ'zߧ_1| 6ovS7Zw-k(źWr m_n;rhM 0GG|)M!,%}KW Yvy~+62HCl*EŬ崀 1ĦOm6gQFa$xMro^2OO)‡O7uq*JcU H㘉dp*4o~MUY?U!E,{go߬I.'ޮsC k2d| D*sbrBVM̔- W bc+lb w2=5Ϛ62OayPHk 3Ԟk|)kFy/mEE#Oy6 jh<:}=#? *w"ۼ86pGT`e ӽ:0&^QhП9PyN%=*w>^DCD#l֦yU>^?,7pX>+n \Q՗?a$LG /9qd#(GK%sI(`)FSyKs·yfjCeH02Y 2$+Z6SJԹxY\nh9fgpE3 tP:VRД@SÎe*ڃRK}ݭ;9-?qF[Wtτx޾g>3;3zPΆnngM0g9x`bR?*2\:,?.y=Eo>,bXB4(_GNS{ҢP+Pb*tc-oP*W .OzQ 17`&$XCeOOL'Ψ*&.S(3<%2+63єO(>m|<NgCI̽nDO3#65ĉYSYZ{ҪHϺ gvOSI$ YK!hPg'AɯuBH9Ūjԭ4۷Ŏy uzQݝƟ B|p)_;ʰ0msO~%JuynVxZ1}j\j5{Ɨ<;#ܿŻLV\?4mAp73cMNGj.~1G512.%mOs=ۆv®n;CLR.4UKMT/t>-nrgC-2lP~, ̴m >Aq4~n]-BְJ|>}][9v+_ Rj/i fB]Xv6vf=nJ*%UVa;]Z(^"~IJ\ ҏ>Z(^cNewUD 'E/gV/gV/gV/gUfw䘔:]IeY*Fdʬ`DR TLhȀ%(߁;|ic8m<]bq)(+ Xgh X RdfR)*s#ys|J;0{иk]; ?} \"{+qE0|Um-ʜ QPڍ#eLri^ JqXb9+%b:OVB.-zY"X{қXDAI.,9ͅJF(ϑ2K)Ul2Kаe,@M$u/?ޜ\ٟ"īcHuP q<ȆfҊEf HVۚ0^mX"A%56jg'q=$E{*c:}Ek6l MҋfOdzXz8^ݤ_抰Y~2y$`{oj)*Mq *N}Zڵvu7<5ViRZZER2|\&y:K/a+O6"t}?ꊜxG-iR*)䏓z\z&^׻by:N~w.?Xot; {Nx?o,J6+cmW[Ŭ=</WܫV3Z{ŋ>e'\h׿}fCnm[Wz)Y7Zǎ-\z?qY7Z Ѻb:cn}";n][ r=[yRjFH:e ƻ[uRGjyncjVʯ~Do@:yEy;K4pRԔb#xw2B_啈Qw֪̭pS{ 4y?}^Nlcdnf3Io6cו9NM sl=z14̰݈1! r3& acA>r|=?9ộ ]=bztgvi ut<с#80e[q' tKT!nI|4D!oS^_Kxe^{6jy~K|r,2 C˜sL|>yI2حpkKqxPgTl ]h6p]L̀^? v7P5l@tP ]?Bتm!{dVa/z +σ߻RsNύ}Ԝ JŐҎmy=}hy{i1D 0ȯaw0v&=q:pwiGm\[g;g- b֟gTw, rOB | I;p/ܽ:* :9A"Eώ*%,rWq%Q,7/41`c5)õq`t[bp%̓|tLƏu%]2:Էe~2Ҭ-;-̒JY+^߯$bÏ?tp:Kt'+ ֋nۻ,_,>,JcXx zHEI8@bO"sG C{eCZ 0C e_ +[RV&8Gdt~JbL-3b߽J]Umϲر7\;qwkE#+ىm$T0mPPJML!4GȢEIGDSUCrHE,yX$F4cc/I*6*MH0sIҴT `9b1i3Kñ^op-~lC9'7f^>zTaV ĝ*pNVܭirڷln<*~3#7Llじ9jLɩ62gs2@Ε#x |-}ڄ"vF ȅ zPWOi?!$:AScRRR +s?V6Ȯs?o-1MVo[ne.ns$Ӌ. Q߿]93rfv̪+g|~,J E) )aƦ@J2/hn&$RBg.(R;VпG|?hK dS|!V+n"Ϻ^l-&rIyin! DzܭruF[bRud91shvIT(*qb"@W Jѡx@K4?< S^??C;Q#qdCW\X E*RSV)nf=t3롛YݬꡫF  SԼiO@Xe)BL`LpZ@yE ^NW%ixOFLdEŐά!YC:tV5U5D sS 2yjԆg͏0ʔԬpVc߁wXKi.Kr}w7F}3v8v\K XejҮOIaCVV1+}$ݴ`j]`2<~>rpL-bż x7m[OYɼuLE쾊,N{`HusL>j.|356(eY0KJV1'Jiud;$6@W p|7In0'Bv?}d`#cg4[uO1r -~_90ƥCښ"0/_*!aai/KBufbJ⢙OrP>x|~xN Z_P쐇A~|S\0эwwţ&ӌ>øS3:AhFoe)NjY_жfS`\'qo'4yJ߽XlQ: C2 +[GF~<[0S͕~tGFJCOV]{#؜6'~OG;Ōè$%0WyA0nr~g. *3b--#$4j*^߿/E懫+r4I#(U@FRcEɈByFTN'zɘٛ! d.-M=|Hm$Rȥo݉FRqD),z:@vaY|=EOmD \@-eM-ҴPx']z Dݻx]2`%zJpJ t}Nq0NS0Q"1xGބ}zmMPi aF&ˋynxp ] CjN! .kNy8F0P6Pdx9N=tH4o)dݤ\ +0T7h )v,֢v># )zm ВUx.x43>c.U ;_, ZE*<-Bq*Ry/!cuEBXuo2$hYU ozp:(@M:ܪ1j !PT JDHIvH8fs@Go`Q>L2$D~1S+BOQ^6R@SM _%;S2HR/Ҷ;., R1S7m&gNըI{Hn%,C*RV#ɪ*0e]ۧCNbČ}v$OXjpKeR+|/ *X%rLD,KP*KJ*!޻o*˷nMp+*80t&Ӱˢ:캴k^'}Y<_/~%p[Ǹ(R$!O9 lpI(;WY?wW~Q\`Y._[/E 鯼tё]SL\Haa$g GiGppoih]Fa8R^FA T) <~uD(ތ0ZClcpwQ mDiX͕G"6[I$|w-6iGr1tuCi i Nw|}fL\bd:MmHR8I<թ$D4Z~2̉&Cq"#t!1. ʑ Džzzq`_9hqq=]'< B]عcwmcY=-Lg_C=i/h`\ZJdIe;~gIYeɢLR,M#Kݙg,\]T7j^|ʆ4me3ƴq}{ĶҠq 73{6h5WBm ̷f6/7aNrܴGa 17:uh*Pr jR zLBIՌƤ[ D=)vso䪢LSwjOvq.HnX=ɭyQ.ZMױ[PVS卼dRo-%̅LU}$8\^ѣm0K8$uAF2u K}{6?O󖺄J @N؅6}m藏6tZݍ-7L.IyɏZ:)VڶiyV)mg֍.BEVP{Z>ƤdA-% p9͓GܯyA*k:-O$j1+VfHb ]ܬE8r( \1q jCb u\eٷ"U,oնwjK!TUZX3Uko_>KY%D Ӣy͏]\־ݢ!k[^Βi&g+b$kX7=Nʉ+]TF x^wr2Zw "7p|:\hT!jI)mtɵNB.9+B6`.KD`c/OJ)s 6P%.›8ΜChR$tĴVױ3:Tn3))Hdew gbO67O%&QR|Rmq#5O8{oQR|dVԱ aRkUNm㿍B-@-.8jµ岶,&Zh][*qմi O<еd1êlkiHr 6>MV/[B)UgƚTz`S+Ooص;6YI@6KK kɺ&)KkTM+#l$5x/B߷xl3!qJ6_6x_ZڲP\M^ۭP3xėp+j+H+Ŗm4ٕ )cUlQP#IhԾU;FD7Zז~dYfjgbn`OOn[aI4r+`ޔLyu ^Wڷj/jGGʇ4dE]ԙ"[WcQ6~{}}[\@7lt0:I% `eӟR&?Fnǎ"yyiAe v4O:ņSm_?-w=3svQԏ -NG ;NF k&(\Z [9mөiD2axZu^{fiǤWiSp)uzYhW`;095NIl{yp37f"?$w"Ź1<,N& IZM³ RD7[dꆁ w>9.G9t?O/pb8 kF$`ڝW? z|rF^Mps|^;6^,95|?|ﲴü4tsYFlYNM'Wo͵pM?ʼ>d+]@*;/OP SAV'´\$uY;I+`#<.HmQ.Lqr-Ryϗdf ,&InSQ%~NA2=tY_~dGi_(9>rϦ7׼YLwǣA^vĿ@T8^*0\/3'aK36m`?䏮^D`?ۏ.`۞(Cۻ&l"p| C_GѴSk%̎ F^'2 q~+?N6UPz'E ^b1).Ǯ>\l÷6|}Shr)ɖ~Tbvo7f/=f2ugZ|6|c5+_9qu%s)Jݬ\"3NY|ZŤu14Bp^X.20pNNaʐÚJcyv`k=!?)%>.SJE~PZG'Z/9$_1X D`*ON\\.Oia]d x1oز(,- .Tc` X5`<!b-{!xeU}^${8{ 2'q[8?_c2Lr/U(69/DhclEhĜSg&һA1G?y_D%jy H[~4x7C(>Cp^D=3s sO|؋tOV= (.W Ur)zJ("{/-ntK+B6m0qKk#(0XKfR3+B"OK/L"`&<"=2y&`.+uK'ncu ,Y)j|rnML_gf 2U>Bo ?0;IVᇱJfaZ,r!ZG7bz:ǃ0;E;tsPJIR["Af 9) #il5ĬEN7 {} -nIAU5'Jl p$EEoi^}ˏg;~nx~uAD }H44nx$>8<ߏ[l2SW6Ik1Mi A)EEX* VR\h(=^2 {3?oWU?W1\7\||` IjfJhB+.Yqϊ Yql\< ΁{}c )F+:}ǣI)9Ak#%< q\ 6, m#9З]@nwUl/6/ n[mWMR҈3heQ=5^UO6zy {%h #0XI+D%T8>q ]4JqP%HpñjɅbak*"%lx Dvx\C0R\J&lsiceB 0^NNB{w8&ͷ'.k_pN)ńS2XNKm$9R.JTbj(HZ FpրBS(%W%+"i֞ƭ]ԩuW]tpTm]#N2mCk%8Gdm&z"M٭)Gb(7 ȵ9UM$5W4 s):r&yDJH .(q2ι@:Y[ׇ7׏?뇭FǍǧS^#s*W7. jwDu= =*8vպyN8DpZyS rE dɏsg)Rܺ>uu}ުkgi{czg;& lW9I}rnr |uմNS0}anQl[n~cTȼE3}?v!|0[,/) /ͱwf_]p>|7(xp>^34ZLzї .o#"BbtU߯:H7UnVk|,+ROI-r-Ųx[T{e:cKM:e@8_+N#FEb"F[kc9ԓ|[0zs=ٜ86JM.,sʴV{-j1KukHfB"(osrFn2N ޵(jXEik0C6y?/vUR>NJ#~e,-O^o/*Kuĕw7}9]۟V9m Hk)6g9ah0XM~2ZMxaqSFLcfp4 xp 6"7A[^ݒΝpt{CXe.qU(zΗIG(ta%#׏tt JSGSOԛ9Քz7R_Ӕz*ngH]gGڰ7TkQTwNUQVfU]߬su.ެxIT^Gg,)/&/nl_{@4xx>kO_wQs]vV1?M]4䕫h'Onޭ#Ngn/gpޭ0jSփr_m{G9_;QTy`pp QWBFHZnvǦ-A"=D~gT)#--qt 20QT {xTΈSB Z R eP6` ._xJoh*-E(bWhn,P Xe]-9`}y(w<RBHmx]Di@ .|ס!pӛN:mLQ+] )qP$& eYM ,0?eL3GnE Y0\)!5] U"kn\-9=8ZOI";VT6@M,xi,r\E TH)v0?PaZIFPײ+ihIԲghA'0mu}|JLz*idW+]Ӧ]2鸲$ֳPMvڥ8n3P)nѳ҂wpAr9HBQfҌ?Sa9~$@BVZ(I&KY!|HwuiX1#nqvTRoFjn{YAS}NE $%QN\±bJiDu#=LtvIikQ>%g:z% ?gI9rPRN,Ӗ "YW: 85,PYI A`}Q撂r˜ϣ]I)rF|@s> @B* VBP"9_8.*IʵtD6ڜ%5˕ Qh¢L#s+{JhG}r3 |-E͋9qA9 m}`oݢ og>7O{Oylkx;-XB E49(&+lPq"1+x[UCN&=k"$ 遏^x1xe @ oug҂c$"8z:dФh #X0eJё1 SdMYj@8)RX/}ptAY1Auhr"vVm]bHD{Ƚ`I(\ 4ML,qY_punE8 `iFRj K𑽵tp4W8JJӮj,k[MXКg?Z(+tיu&+Qia c0#v4W4'qVRCkE[qvMkFg =KR#ןcq5nUۛ\!w5ښXjH,elQ$z iaUcߌ<kkm~Vi~_zVgLל-'W9ycqyytHwOP"-v_rkd6KC=+;whT`,HU$+Η`*SI(OL{ eؠj c.>˜RQQ8& !H'#mb>g,&/5!ᄱ/-|E$Q7Z5vZl>MoyƽS*eG j.|ǹt HÉ!NĻl|g7{+59N/)@ $VԳ|y?N-!7ZHE#qMhGVZa89v+;}fg;To3T9~+|i~U96心}j%mmp|*44 ,>DCiXq3/VY!Orޟ'a`:cVOΛƷ?Oht2~6oо?qښur`-0C[;@5$SMF%tڔ$b %xvF HvBk婌B:!3v]jF4$)4 P'ƚ&9 R]&xnD\/?Klyi JjW`@hՁ \*Jn3]rhYf 1+dmgOӂNx0y;a&,%Z*![@4"F(euQ:z/T(Ȑ(e oPCkpoqD(Gm=HhdbH1y4 b]Emj}IQZS `]#ێN/&߂^`njfJ ap> -I f5ՁvIףX- MEZII1οrPwQ5Rc4,>WoD"sQ|;B4sE IvP9Ru& "q0d^>ƴw5LL5n5+ d{q_uFif6_ꆂ sad J ƦR!k -mΨvSB[ϐu29~܎qKr}_\^- m\^ 1ZTT3e*l6]|w}4w=Ns4w=N]ʻ˛w) fCetX>Tx%i)DHboFn'z3ps#.ο{limxh9Ve8[/hܜLNv Ds`݁&@*igaW=+[O/ԅl=$=D09cx;uS f-!']B.: ^;М - 8z7&OBʱ3N`29|YWٺkI,3HxH1AM'˝|F[Os|<ApVaX#?z̶GjވuG-k=Ճ_{7ɏ&o#ib| gGp džz 炿h\^>r]P#=wz}1Xo2_+ˎuFQ}BF;aw,ځg+knF]o,)܇"4ݞpζ=y6%$#7QE%(AҢíėD&pҟю'"QT1 H,ahsii,[šeqT"}61j@ $8rW9ICcAlvL"0נA!)U2;@1mdZgL [?cvI][;Szs<Ҵ%z)\)g0 +͂5A -&H %sS#LW%gDL d2ьP)!b8xՁ8 K-qV: WXf.%V$!f3o 0xLH f2y[}B)8my@P9A~sB1t8Cf3M4sRrF<#& Cf: !Al㑘\*A$OĔH?1)zxewO"`P9WH^d((<%bSTUߦ%L)O~,::#o4]'{[b.LP6n`}p=QWgWbP ?I9i n=EXz Nd{V@x3y;yAI'.Dx=ՠê=}>,ـ6t4'xFRуͭ*ܹݥEGUC%PKUY+uCyLrtbʛ[yz~qzR5\ R,TwEZ2 MĹțHrwK43̇*Af}^3/"jq-|Jm'v^ۼkt~@CTeHVL>Ww22\>>ݕ!Y70b;yaUd31`*|DbRCmD7~zf:~|`rs|@20l #w ;pb=Pw3BWŔ|6ö!Lݷoo UF:\ ؕRޱ-Hj!:dNv^xVʐT.j.v {^x ?$ly+J#-6-ha5K1'*..0]+hEh;/4]vKW-/uɍv"O(eGY|U1a7w#KrSėߞ?%8;Wp)Z&ED*ʮ vCW;+u(,K|V?kRҙ5{*t1\\f+T][KMQ,;!+ޒ2H+"Ua{K [[Nd:#{SVs [qʐKD:iv==m tR]c;=P%-4=旭!uHPuJ&*~&bIʑXRN%*hF)S|nUjN2/?)q$(ݙ=Mu^tC yjhILo<:f=mO#{bVJI1f5/꾣;ޅitUpiV:h}j{u9ƩJbw2jb2EԥkRL 1qL ì?H|UZ(!*k Ƚc/A@C4%>ADb.]u@5mP +2<*`2|'))!}ƃNqR뫹8>j i!㊿}OZ61ƭ"^G::*p{V!Fhۅk5fFckFϒ#/w񕢅YT'E%f|}μ( dъTZvwu @yg{\Ya&lG>/$ge$RC? pL97lh\;%3[Ī RZK?ë^4AHzC˽\SŒ'n>߀/q^?̖^׏U⪤"ݠ1ƺ=zm@)p^;\H.y Uֈ/IFȹf-AZ3wSTr)تc\[=_._e[ UƵDjXeLw>BYw-\*~Sq}WH+WW=Mpͥ|-/ɟj~cˉOfa,"z+x[P&F!'z^k#t?jPhM̏#] Ց\Ѽ٭9Y)n2mrc%l{9'擐?ND&J7[Օ{ '/6$ Eqn6Ur)pͺ P NT?;7HݝnrsIY+wYsShg Dԛ4-N>^YoySv㬕$)5z֚C6OU"(„FȯS(Wb$%[z, fw&`&`%D,Ht@bu%h^$%CR#*YU\lv+܂Cr!kq}3b}WW۫<zח??L@%R QFG߭a%=&Dá}ԤRdɿ&$׫\˵&z@9Cı 39*.Vg2bjUܻB"Ͼy/=y?5y1熕*#g3 coݫ''DxJӅpAzyleF%Z J2{ @ϙrȀP"׷ΫGQ jX֐|6oN;SR}lb@r`Ŕ+Emq{ ̀g 8,!7üT+vgwÃn616`K ~¡͟=ω<?pKTL1wϯJL&ހŏ163W+;z9C0LgsE|4oS32X}Gd+Aw0 aB\>XFi!dGճ#G(z쨚ȥJ*`V 0iRrsvD.Qh-bxW@#}Lzt/ekAw9),핿ǡ6.s{uew{PCfjuDQSp'H,ZOqhT1G+3C@b"M#E;jOh/IW[P/-hv$M!]\ۡ)!V-W\VU(-9H>]|i%KR,q!$ǘm#ZtIȜ3Y-Qm-I\d|8 -ܐ$U9aŀb5:Q"*!$jCwB0aT^l̷w7:sLGCgq0O.wq]y޸/eS`kkZ;C)H Ƹ[5;9 i&8s.+C f0%΁/}|{  I,x D 7K/2} ]nʊ".8k/SmWA -xCCƅ"!15yW"!Fɂ^r"BR!is}61̧`\mxUxϫmoڻF\ Ϛ0eS~=>Ј P P}Tz{4Vjb F^۰&'b%&NXؓ˞]ݧ_[HG4k-R.4C¡Q,Y[g10mn`o7j:AP\ 2BF,qۢm]3+b> "gTMDtRR$3h53ڎjeT\ 070;]i8ʬ,-VT7Eժ绀8\O7օpPoٗY x?~[Ӎ]u(@LI9}FV[;ҭ%M8uJ㉚8M(oߡU콜ioOq?;SMuJ:8(tFLm QH?6Ƚs )]!i8M#|`ф4A +y%E6R÷dxP9^Id՝) 7[#N4d<%($a~dh7IOh8L9wAׂhj4(@G%.;j45w9!)?u z wrF'V2@3*4q* A{`{O~j|;-1p0Ǜ>|wgeL*؃i=[Rk{Ζ> pQh4rpRVL4A;ã%TD?.L%qbA㈂=c<_هRRϻh"R>CߙL\P0СY#sEۛa,OQٌNVif:^f$+8zMt<ߺ~.TVTw퀪"0$Ҥtfvw;sv+^II?F)rl.A+ɔzj0:GcSW6588.'eR<>95 ߎԉ1 4;5wm׳yQ=#\@@/6)ٛ/wB"!̖Cʃ+ۺrгmo>[~XO3>YB%A˾Y:Od'T:娑Ǥvnq:ƹxo ?ڇ)L1$GUDsq"j&DsE^Hl\S(fƍpKYiT~xtdiJdiJ֓(Yqg)1$f; pB=MdEךEWt[T~qҁӕ^>fTH>Gt?MGt?o/2fRHߓR*bUnZp Lr&99P*"(%# 737˹J:,]% u/^X&`^ZyB(b2H (k`ׁPf(alЬK ;QY酛J8*Wlgl:I[Mងf0gLJEc+` HQ %x?CݝOwl*/9-+xDUqEwҁګ;ή'߽zEњ[')ަThה21YLZ9 !!Xi-.~\L5%O9mk/uޣUFݞzLR01 `Jp3AC?庡 ;AgtDz6'Ƞ W9nByB[WWBG#ȩlJN1%N$'eTDgD+t-TѐHApʕTsS;Bd#ɞBm"Z"J8nԳND)H glpAj* @o%$0 4DPT#Q2dNDAH7P!ږDΙj64P=9> `q) Jshfv!@=ܒRH˒B;ҐA ^0[O(Qj,&b.?+8 <<}P,<8ӻ/ =۩ O3O]L{|9/O|f_=ܾl$ 3=~dS{3t+m?RY?ل1BFY p[joڇq A|=^"Р޿q>|퓅1F 9F}Zlc퓗>YX0L}&|zpSb7] 6a|8ٲ9|z幛9BtgP7 ќlmG PˠOoHӍ L 'Ne% 3Zpu\C$|D V9nMx zAC)x/142@FSH434meDWN|O'@{CgϲP)ˤ]&fӥAU 6~紷׳^~}^\WVCJʺ}dJ1J247O>dO#XV͡5U[{X^D#}q>Okչ%O+݃I=7Kmʕ[;8D8ڽhJplv7(dF|~.EdB;إ:eW*K]*;1fcԑEdQqBx [TrB; HsV;we<!;KV[!X+kBTTg9ƥ [@0MBDLeܠEzYYk]1UkGONA]śV/\hD||c!2HD"o/t4Jl!?T ٦k*נ޻_`k MuLGIew5{K-$Ւ3KB_QLJNv"7 q"!Nv~ ZX;Lu$#ۗ_\ge'l8DN{ihfm8;P܄nEoF|Y?^VxvEt1F`GhљmysǓRصv@ m= ngʞ?9:Ɵ}ڽ\P@̈n0&\%d=j.d黃;*9TZv#{l6$=cuV;j:,S_\k3v /|t@Ytïoe; =cr`W{g>9h&{XCyAP#N޿h#:KKz = 2)y+u[d=0ۊY!W˧_`9 zl.\I& 3sۙӃo#A}ގH8BHW+$IkWH t=}Iy+WHXE=m_q8)/-@ڥY-Re~Cd_m韔iX0hQK+}5A;!DdO&u4E1LS4E1LQ 7hO3V*BŜa\xUQL)U)n=h픫$ WvKy;.f~E@]\#gYXc^<|A\^[b̩cOq#Ts"Y!h"+%X\B\%5QYIoų+Rzq'Mf>=6E<:XF9jʘ _&eMьZp 2[QF2jd2޼)G cj4DJF찲qЎ- _) }T#}yF~1/Y@1+{ "TqEkX;Rδ)REfDIO&E<7˵\Pψc,|Zpp;AF cClǚū1ݧͦX͙'myPpvSpRTyr5bjd|>H[V$U^)C(LAR7/rd/SA5=N)gHN Hm [Z q Ϣ)vA#( -\S2E"cy"$ #F)@PSYybt/h/@L (rMD Me6l]2TNr)DIm@x+^*:cBR {^ꭊh:>Lfw'P; ;t{\ a절ž4ܹߦ3گ,}}|ӫ񣟮o| !LMOf6#I_ކnr_d8gf6u2A㉘Y8& lH̬ʬ˥#t}ھjkB_w/K)}nק[|> $*[)_^\~O)FY(ND6alhKI|F'8r< \ˋw0C'5}n YT * G_͇q~2s,W9M%xAe,AUkw.Fe?NJ5c vZd[4 &S.!IQvnk4=$iVpxڵ}>5:{('aF++ס fKbjXrm,{|8»nv@2`d8XmFWlk>{N1"]gC1MYwhTK98V==6׾斧;{UBWd2ͬcQ5<:a50q)P]ko6zʋj+(z?z2F~x6UކPWyꪾ G'@}PYM@p%F!`B=w$pfI!PCz:~|o? NVW~ybtW~[^G].i߾!VZ .HgLFYz'uImHpa:X@wu~p0̩4;#H Dftygh%n ,x>Q!(A<j)ܻNVVQH6E5­tETe cօ)e8?޲)(xpfV\] 7V;&@x+{׺OQe^|ڨP m۲)޲\<=2k˲\ *{ϳGXZ& DI9_ m_^M2ܻN(#pS 3D3yryGFGpR)9O3jRŒ%>8y"nrS85,340k/I1r !BܔfEkFS`\\ڎ ZVfT !]h2D|fx.bskk3cT딲`AB\LFׄXCXQ[.:`@Fڲ : ؏6ZZ~Ӈ>Qipfeuy~rpy[>8{\e=nڇw_ܻň^*ˣpb |eTS%8ip@NI(F2$'>z)}0\h A{)OJZM9N@ ˣ y:kZ͹6"vEe)i?D(AT"p2,JX E k:ڛХ|;Q)8 kN(qEW^~ SMdpe2[7BV&]sS# icçO?Md)nknKɗ仜.wF{<寗/V\rӻJ9vAh?1lw&M~R-$NS!\PIY-~۴W}Ik?D?S,A9-)BړԒZ":L5 oѹoG1J(*Tw1^ݥ_fU^/*Sԫ%:ӷzPͼoI-bۋԲqPҮ̞[ԣ$O=6/{] @MVQHH9pC?sZÀ8dE:Pd BB̰~pu!9&I<j7 IPZʁiNN[?{LgΔΞ.Fz!-UFY<_HAtn!8*(!k+_-9\%pKuqHAArJh7K @HDp:*!EL%*<1Vb7S<~DYzG֫<^Gֺ}SҐX F;ϙJ +A+rN{u(=W|^e|7u kѦMJRTX[qLѕ3"$0Xh+gB,GqF@рfTRʨcʳcʣBGqjL-&D'%*EŘڑDMd NJ嬋$Bt7$ A#o;!(lw’g?J!FHp qA :4V9M<[.,n**N®E'DXG6WRbjjhg*Eh2+k)#SAE]F_ cqbH(hM6!D2+$Yˀzm>J{npa~=dQ zB`CO&E-(6#ک+x0VA@tRD&st"oW-LͿZ>= O0뗿~\y SF%F1JMqYPϼo_n057@@V^;|O23T~~̩6eZ )9s1ǿ,HE(k=p8+-/&wKp?=b[c0&ڐOf8K(<X ަF1Z`m1CZ[Adz|.Zt(<`;-J !?i[h,kеSukzQE^q]7B[Sj;j%~-#zLޅPA9v_PlhTwYm19fai|w#2Bd¼Ԕ24 )'8 '<)|5-m0ơĨQN5Lȳq(03L 񆱡\r6IÄDli$lakhO1cMFT&1m$)26mmncF+; n2ѫ19{ l8!_ik JAOn5}i+mv_ΫX"Nn n?)@b`K:Fj@VkD!Ҩ(4'WSZr6P?m *JtdUr9cLRfatM'0i}NҢZ t#q.UxA1ʨRNQm.E*n1-WxkG 9ξG5jr;^9OR.>Զe :Eo+0[%pK==B=-s7uz Z}:S@XJǣZĠYԯ 佝CۭN [. F.͆;[ոRe{jK%=MBj q.'qF;#3PdB}%k~zt=6FQY26F_c.ѧf1רZI9?^`Zw*T_=w=P޿󌺒>$8P.׼A"5S?Mv1N?L UX>?ɵq&_nK^47v]78FF:I|!,stNxL'ia'0N0K%Y_\VV-QvLT ʣ_iatP!Ф8f0%UQC?H%> K~36$R]-L̸6X=O9;IRʼlM'gh0zwŽ[~]^X@rmJ92łSd4(+ER  s #sĆLjyjyh䎑!񤣳VOm$@ T&*r7gBN&'wAIÐ!0K@5aH X3'Y&GdHdG  C[\CYHC+2}mWy$',$GaP J睍VgZ]]R4d](r*pJD E1IRƕ'""NSk<`aX)D=:m=%Q5^0Mհ7 l$"S!xoՐsYD_ZS3)'{ rԱ&P+'*Sd ^ZZShpƥu k W./MaoS).g{[\S p;+R2UHސúBU|Ӽ)}hE߼m^4҇f'C5TγZ36Qz ߷p;OvsyG qN=h=Rs[HxתT< ?`oMĹtϹWKM\gM?{W7r /6:']τglx.H8oVwDF#DA eV^Uy|^ F3\$(=f0KK&RRp)kTLDxeWY-,Z蚂s\R"y@K 0VN-̂lWU[1@|tcfІ HNi_ik!r!Xk->Asa\)T RQςQa{B:[@,( 8CezҐކ1 -hiCce`ä?-FG=^?돏x_b(|!`qf !L۝, ae1cݘjzAVX[4FWRzI't"Ԙqu@ԱLiέ TDRGb[A L"G.h,vbԤȬLEńz\oQrqW@vm#պ?b97hdytnb7a^OpIt ۏkGCV+ `琟z3 93޾Pl외y<i11M_ psPkrv\OFG/pc[hCUh#/\:8#0T:.V{U{TAʇ? 㿵7%;^1Ѧ-Ql&%i*᥹T{o-0oUpmJyN`Yp[˃sʯAf Zv/~. evS&x;2R[/̈́'!͈F~s\szsjjq%7ofQb%g>w^6aFp;hreŕxZ)7͞ŭ)2'GWF^0 \#"aJTR0`斀l$Zs\jbmxζkmvŒe CmxUr+Cœ  R>cI)=7/A#)mEqנGl6H-rXZ &+7jmN^ s\ ݊& @p҄USIk ^*e{n?b5 wW TfrCK1pH7( rPE`/ΈRWt7P.&훞u% Vv>c\S{7Jwe;yyY+t{9BP H3\ulvyS޹ySŹ320(9Y'i\lp.=y`}TV~}S+VtuvM[N><0u(z:eSB51舍]۳ۋb*62ͧn/YI1ielz|ĉahڭ<[v?["b\6Pr&P-%}o%o" VlN b1@ʻW8:srW`XT+ҋ kv(aڠ<8I"9NOM:_￝6 Eu<9>Y?O.7X?OQ^-ou~+~9OD)}hyYn4Y1rԚ8,VILRcor# y"Z$S+MFW=jR rDmc7)[dLօr}LqjR*%2dKPLod CΆ̷!d-9zoCgw[xwkt7RONҩI:;Iv'S;ࢺ@LPGFlWJ-EG☮%POh,7#{8Gzؿ%O>nhz{27 %6c|?e?k`UökAB~-q3VϿ?2\:) ~QhCta~v4"9CE>֝uxz9ђMQKrhzr@0 {Uw.K,E(*85pOVoXyFUw-3t&ͻ >XN_6<Ùj7ڑ9u_]j*f2*>ƺjPw+v3M׭cbg7mj)t9I| /}tY"o mj lar@>.wHCG}Hӌk_LrO|9@9w'IA^CpkO*_0VU֬6ʹ nuU(N^CrV:F!YSPRt"+zwFi/}Ɋ#@ˉx6dY=."%S؜]B'p+Ϣy+}#>oqM>&K4"@(x NLwV ӏbhҰE2N>lQ~%&|Rg_8+ŜP9A|e5 Ƭ?c)7ܡny }a/ѥ_Q8`!DpYzlgb+QaUkZ.x9̣Mrj}Kέ&rA4[,ߪed39ZG;q1/_@B*ʡow5Fw?Lg I:M3gIF¤9r\eD7 SV@\:Jߎd|C.ԕgVB;QYjYEP8.ҪT}2rbb+%7q B.t̟/Ȉތ1YrPcf|'(I;oR۞ysvsfGKjj6 ͦK! Ws,{PsdMjwm:زҔf\Z)3XF̌$ɫzi }Lq憒d<_Jz2Jҳp\;`l8F.AlMEԦh%Gw+VFIkF!9iWGn9d${Xy\E|4ґ%߃]v<2'/3AXt'/>SOR1$SOR1YLݜB:QHע(x=cv^:kumٓKT4yK%X2M@ Q>*MRoقb΍уOhA`oՂ@)*uU}ڊIڊIiiB@c@q|rXUh3Ý >2ִ*Ҵ|7F,<9IIJoNRzsLo6CՁ[$ . ,B!3I:$v HKBga p gqPiNA(H#I%ՌPC9Au0Ie)%!ZkV Hc@RrLu5zEp :tʜKN ޡTցhւFaUMALc$m Yjʠd+T@)HEreiA`cE-(t/`Q׀0\D_HQid n2s-3_2D= (i4""7"\K9`^MeVH Y2 &_ \ (A).!t`A/ id ~DE/J5 {oyUQfLPP~ q:75$ߑ ЀnfYVe)ůHp gʊ+)z{"H攑 iVNUm Xi$:kB1GS|@  ^]5q .dMR- $Qk)$ZZ$ԚːĀ Ie2~*+֦:2|{OHDv/|xGh0l?~}w[Oď~<FAEXhn,SV?'i̅vnrz_D:ڒ #Þ(gTal)@Y{>Z&Xj8PtR['7,Q$ * rӸK 4kx@S\ 3kn$Hp8:|&I fA}3JױMh:1bdat;A )zQGeHƙYr(@bM9+,G&S D$B-ř1p+d[;τAQ6%SrRKTF?r(E<63|%.JNq ߍ(e̡FyRPލڲ_2f:$%6 8[dr6՗mWQnFMP²1pG>Z@g7a締*g޶B?4k?vuy3y'X9״76Kt|@S΁^hoj}Af! =٣~Ʀ>qCwj5|jNx2џ^YBov+ea g#_=83דO$yP_aw἞[4m̲3!O'kqV/]@ݼ``6'y Bu;c{lev˶dSK[H̨bUL- &zSGFjnHYMEl>Y?Ob{7u3ph NOH(T\eλӬn@DE޴6P6ONf((D:S oz*DqZ&7p(@ʪ(dkv[ 5%Jv꿓 7o܉ojCd,gwFcs.[FIΑƣk48r0XrF&ً"W_?]g4d, 1}?D2_H(b}/"T[s;?z y)!|9}>,PILѾ CdWf%d&l"̤@ɜ={u"d8QPIT)Pi[c=:g [q,#*V (m&JN"HQL#HPfQcz3CX"Mn(lEiCY|8-:=v,+XAI(uӕDg5>E[ӱhrM"*vҮ2R2sx|́D<u+ZkI^OC6['C6?@cB}ԚI$W"VӺ=*tC尊]opaJ-:Sw^߲=l3 l12V_YG6OpgT!l1k֗g0*)<pSЌxŒMexJ~Xxf%3>њuѴjX b#Y{~zMr -jap{C<)U/hŤ*]|&Wzw ImϢWHDw Η&OHxCCW̪۱>ٍ6ڞڇyuX|(`Z>X3yY,+F% A39߈l2[|6?;=~|+>$z^&|fӌ{@ys1n/QbIVX|+$r~ly_U %WuDb|btF>lޞlV۞6b cDsTHK)umʡR$IHJYzizp<}iRuyHyɛwB"BVg7 _bv|pP|^`UWpDsp;@1}XIxi=<Ԯk#4eWG`4QR.cev̹$0a*ˈD *4112qUpxti UT`};mJ)Q[#! _ƀi”c$%TJ.f*:F\3w `j&F#pi\dri~4x`~'a$ 42#aasi{ S~iRg-9,M%<4 3'BAcC$dMhdfg6& Ԍj)#BHIlQ:JC5ODJh3(h U10g]KJBcX!@!bPP$eDF)AJ(C\}4WC{2uWMS5'PMV rMg`4fGGNGFPʜ*ƪn2BM8 ~ԾZmfE5SN5E/%L6m$+6Q bwKq^}ݕ@>8`x-*}fUx_t^%EsNo[>o>JreɸIH+[k_a76KEVcaoO9J%hFK$U;J-:(cPSPqDfS}.ޱƂZLoi6/ȥ=~VZiejE3c>fw7 9X7q չ۬Q%v֟WOoSa6m*WbJ(1wFGNd X B΅pdn逴;a ,قcb"}bnI5L$vMk+8r}-P"OT b$4R,ѡ-X !L%ILFTK%+o:&#Oo{Bxd&{n?&cBG|߆!V&*?gdʹX.'wQ ";:Ȯw I )h(-4O+dC>ڐ|v}Xic}|t4rPT1ߌ\5zܴ=RpI]Vdш!]eP:؈\\ŭTe9®\{+'禩GN*K9ˮoBktƊ_yש{1Yc!5q7~i ssbp8}P;&5UC3uy3%K=qc1[RqPix-5jM|ԇb*ԕ}MD|8x|H><J-9/?Nu Qa%/. OYn&_u*]<̮,Kz}{| 1&Q&}Hd#>eko ĽʘOh bn>dS_=7Sv+xa~b~"ŋWT5Xʵ1Cevo@n,}JC$BJ>s$-^Αlg\wGFUV\V,h{T-eEIYRhI7l*Lv>#!q62IW1`  &N",:rQ, YJ/r=L3$oskCfBwGSWsC1z`OUɳMugryZ~JiVOOm ○Wǟno17^_?]%pzX_%3 3iݲ\ϳ!5pS9 !E~_/fg,p Snbdf :lzi[aOSĖ6ͅ@sfE$\58(()Ǚy+{D8kXV6P$4;ٿD:;)A:4ˉei=ҽg2E \wͺq \)Kz(\sAj ÀFC}V0iC?N<}F$z2Gu,ǩt+|k1xj2ZTky((X2Ac 򜴷5w .6_%M\+ZJ1T0gʼnu˜RC|,2:NmQ">nƧ} > j0u8klۥ.f5gK7uA +({x]0Rqym$ {]_l*WT4D4헵-|<+.FG_WzU#A6KxCm64wk#nRDu0SD\)5"Y#"IcM(" m^Zm2*!/  B/pr)n3l6Kܝ]W߃dWwMHg`a*́h0/7sLƲ5,Ǜq춵E9 :}hK*.1K R| g> pPsx'sژd{~||]68ܷ R,tm ofk$' r6q&|J~w5 D626(\D3w3jFˉ/1 v[AN*_oR;?㎊dl458φvl]a-z໾<ϩA^"_1Ȟʺ P^VAj[; gj"4n6="㞿1"WoOUDuv~I.\(ӓ̀PcP,}LړYGeWRܳN +|t픥q:,2Z|W 6X"6RMa&IHsġJRT$t؟V"( aEʩ8 Fn4T,`(EA#D1Ac`AvZzqHd9K+$6pQRtG͏eYaQW%Ƃ6o²^pbD a~7>Dm bKLw<(j;nIg֒x,g3G=M- )XA(0 85;V1J/%涝 6-NTu^6苳$ۿvuICd.o5:AV9oҊcV{mq/^ϽVztH{b0mWdtxLyUW|5?Zo]$Pjǎ9ٽudOGնRRsw P)Z^ /t,+;=:^_a.[E^Kzy]joY鯒o V6}֮kEU-a"(o> v羛lF: @ e֤L)?rRKE;-|٥^Zٕ $dޯ!Z8##aǧh:3RyM]\n9KL%(_Iʣ^ź-(qzk=Y^ @ڌ6= m:>kD$&F0 (`XHISYPr:bBZ)f(kq-+*"Ÿ| cG5+FScEXw.+"CNj\PK_JA3ĖgòzoQK1kȞQ_!V~6e.~XQ͵=UQ30ʕ"/WgLZJFD9E?j3KLh7׼/# !>>Z8]e dթ]1k >d^ cgkmH99#uW"Fl^6 jq%Q2IY?դ$o Ë-q]Uuݺ̬ n^RkτBńF` $/LƅHX2{^l(1\ R?BfJ <#я^Hk||e2&l1ꖼҪ(X"`J+1}ٹ3Yى;~Yw]-#=Vjwjo288+koi3.;۬yLxm[;W@޵ rx>]R_$4~z>.O&r_sgGݳ=Pϓ1FJdVd5y8((Մ ֿvt볁1(-\a(]2mË(=csiIUoW˯ I$#·}eHM%Yg_vũf{9j,eW#[3R:Thux}m40)xWl $'ub0%")[ ,HGE ~0jB,S z)t\ Դ3zF h9m.FY -)f Wnl},P*D;9|#ى?C?9=pNs0^?;hKF>nDO#;v8FՆNφF9;XkotƓ.8j1LRL X/@~Bd#^.9zL. 7Ʉq*E6*Lw]%, pHKg1Tbk[ 0ûϮzUHZE\Q8,aYd'V洕UeYnm$"C1+VZGJbb9^1B=8FoVyu8$oT7:dclh"!bL(|rCLW/2$AEA,-!BG/(O!q@)aL0IJuw'!KhTNC4Q+/e|yUs8ގ#xzkcR(V1ŵމ x7u[1[GPաqFv"vYfٔ8bulxkͼ PTFC ch:8SI#8w\W8#|ہ(4Qvp#CDV;;XI4N^?d_A*@l}=KǓBϮ4!'=fkM]pN1ɼ"om7:mᴐc>giA^)X~?|cc{Kw'߻OOZ':Q։"N։: |I?y;O6r, 1FңȕE(rGQZڈ&QʗuGڋwNL!pjBYbuѽ2ٜ0K^$ #Xx*#H q,-XQ)%Ueԅ)Vdz?̱ hL6gmJY4lRoعB RJODHZѻ/HTW 6)t1tJ QScN]!>켖!6[F- :+mԐ+y`"ɥܲbC7N"[R7Skm*Pp#Sᨀ[!T ;Rt"(3;&@Mp`e$xz5M#SQ`x3>n3>nϤ\BF \[Fq}HEEf% <9 UDJȚN+ HKms%d3Z@ ER27*}b.۠UfYffL9ZrSH I[>wsNE$PE4@ (FYIi_Fwc";|WXQ=DHq~ǧ;~XѼ !?5fߧzc({|xIq/T-! [si!__ޝ GW ~J!?\oe>}ghp@ own.>UK ?*` ]{DF˒1JqK),`DW*"1d8qE! Q41p'g p^(bQ&VM  B=iHHYЏ([12DAkΘ>0lZQ6۾1S{+܋\^ \^5mX9csƂ&PoL鈞#=Dt ?D dZa}9#9!X.2 )bF|BXzi;m d)Zd8 PL׵@\0F+Ҭ]Ra0;r=q#5/ftV~.X9MY9kۇajP ]*9bvTq/dx,5ong2/\W9vr Hi]UH| Q}uaZ ͓xY3ij3,5 ʶP^D`Tzh&6M /~/-_zBNt?ɄУJ|d{ ӫ=i~vyO;?_.yѓv&_ar^Ka&M;K5ƙ63Ǹ18hsL8䆢IyW7Hݘꏏ5 SjFvLٍK""!rCy~)7?q~.&%>^ a Яf[X諗~tCn5qo8996^K{"Zജ5hEsZ *0Ul rv[}/{vRxO =s7YA,Sp&ya`6.D sAN$҉kN]4Fk:M# o+'܆t7-;AFJE8'ʸ)hBIѨX +)$(2Kƃk~g9׃N=@CN*sa!>þFթfTx_0i٭rFr*UD[NCR(͎GD)% 06!7$4_tI#Q i)R45$%6PB"x bϐL bdFZ T=@z1%5ec*WpTY|M ccj}m1/W'SPE:uH?EfNؤ,ﮣ&w(?> _wG+/N>o^)8{{/OI o z   Q$w}\\uGN:[Çk=]^&HIB6jD2eHD z@] S8~,1B &Y48Tf٠׉Νxih֔ki%W K/@PsQϔ/,_T. */(L#rTh.1z"fn7E<WA)ɹ3@~e=0 "c(DJGsvq-sVmF RYGRK[sN82V,JCORk!M@c䌳DT.]O$RN4^@Өe: hoK5ΤM4DMD7,wF2pKO@q)3 k{`"|5/:?B}^0XJa;LNӫSߦOW^~Y~+nw8~ F lYt%.,9Y#Tc\lkDz}/Iuc.U pu]YoȲ+3״{_$db#vf0WS'O0VZh-VS"eX@l6j YB)\8Д;dήḠS.>Vhrҳ@#. iJKϸ0L]h3qABL"9'=Rz FI)ZA=URwJ`s;vft$;n=vUK aaJR;!KJ["O#6vlJ?Bglqbv]gZ$Tc%UPMy6l$Q!~\MY ֿȏل\ڐ~CK|œ!qijy~GiqFނsi>ta0DroB#(RZ׊{ydMqÈ9ŰtZRi66>YL >“Ev\nBE;2h'ѴIlu\QXָL;VU&hũ1@y/+,yƮ2H!ƶz`v:ؿL嵨 K#6:N3ʔa2l^ }X8Ao'n ^ÈGɮi::\O~`  .? G|2]ݛt Ĥ~wބY2V$nSl~o~t lj$Р~0Ҕ4f9>0j &ݠ9.4l >ll7a=1[usY0pԉ]c33S=dH56ih |I4rmoNwû'9?:]8:}{q̯z͗W7n=='^>ǧ'^_=m? cj%aɏ`z}+1iӶۺ5<>aVeԛIf8N;srFmݞ{y GWϾ`} aR'p{ _4`ΐkg'dzwɡ+sGsN5]ӳwz}+ ]V'}o bF|צZ,Kvh>t0 UAKK?NݞڭA я0_EuMdou{$ݰs|$?3}v>v)7S@_C239?iG%i:{{7a9]?eNzNнOmE'Iϟr<pV՟_0uI*|2=0q,cHl_F*;aV4ɋH{CnɟOwnڥ9-huRW}YZzP\'XxHT]⁵ D⚍2b!^MMrX^\2?4K>tIӎAPloMcޔߛ Ly)Kc&}Ly7!!T9aTê0wFaHh@h)/'zoД)o#{Ƚ)?FC D|<`Ky`vу808ݴWqCSNk9f, 22];W)hoMcOo{S~oMyS~BΔ℄ːIc'n\s*|LSF ĻPTK=lTlanr"S:puR߽ 1")s|⦠_da~s[w1 # pτ].p:K1c#l_M 4W}6ՙ_D*tH#$(׮:qԸhoZiZ.F`}j.ڄҎNf kju@;f]|f&1f5Dzz6'*8UߪSl()]1V3fU̴c!'"y ﱐ*#=asWsCh." (:V|A&#SΟ`ѭ=8)E>aK,$(Kb5= IxU- @0 |)̳f*Cjckt8m4˨a#h0Q8a.1uÞc^P͜wAR%j/Bi'MTjtd-۳4C"՚peY4r"$BPKe">lBZ˜z¬}w~#ϊ;K ׮.z2|szXmZ?sos1)rY<[`!ljK8GbQ/̚&0FTxo:y ;pUhoY/W~a]WUh|o,]aQz归Y ;S.Uh_uTWlivLtη-82 ~\ 7A2TM)k JkZMM>!yd&'cò|2ˈ'D.PdH:D)Y?iX:&Q[׏[w!V=#_ܪ@K3-3-hҭ.gHGM {L<>`58 |lL}iTYt#ëM@:Z:`.ȞbT1V'>"z=(%B'8DE= Y3 4\D!UY2}i^XSTt,yt#gF5Q|g/M[gRGX.)k 7xp4΀n б.<4uYrV.H+RH}ˈ<#aUjy$&CPgS&2+u.O}=!͞|zLiǶ$cUM#RF5Zt[@C)Y P V\5[b2JjeQ edz$B:t?P)&l)6+BVP;d{Nضɬ;Tg܇;U}[-O/B S@Uy-@x@01C}^mtǣBƂ1W,恈Usbϋ9i]XqZᢂc/4L7p?%b8~}l=ÎtXKkv%ubs [r۞t6eK\eS8߬Tg"ǽch 1(پxNpTX9xJ6f&nd=Iv*XL= n6ϻd/Wz۫7aV_|xq:]eb_mX+AҤ^ƧEou0*ImM|Kxzk PiQ(] :< o&X曼5ʊ5@=M\.# pEu<+Ą/3ܒԮ ]+l|HsTҙ[nd7x8[#bkRKQgmI-;+0hXG1U7?씼5D+J/, 6z{ h{l%KQg'Jtp%YJ^z-][o۸+gQ@rlQ&h],ڢ(*ob!%ۑur֑ii}ѐReeyӅd!_uݽH ad$E^ * 'v^q 'v4DΩq‰-=WG=.wԸ~Iqn /ڛp 0;WyO[T>Q 5U;%U7D8nwtvȧS QnZ_IR bQud;X8?gbTb`"&Xe.c6QPIucn61\=9/4jHL2gp`'74} a"A CBC/JJ |H5#bq``w(9ޱfRдrD drkYwnU'v@_!& @<ƂZ!FYBJ7xVKFGfKKX)"(0Ej/Qi #"B! ?>ƪ)ZT6_p6`HBcEL`c8$И>8 6_hHJ\ȻZr*I2H"!T`10c%iKV#yC\^^kC\n qk'8 -!o\?wy35χκE7x:muK{eZ)'#08Ua!KjZG / uhDM8dV< ǟ{;N*u2sK^sBx!qQ G$qJR\';uz%:ATkk uJgu2 KYkhxlYSZ.dͩq =VZZY ΦT2nZ*T&Th1R1alEsg ״ߪV]Uw}U%vMld5JdS=.$::eղOKFÑUUZͩL倠2jǺKpGJK-WKcMKiWPXe5wI8'L95 1nZ%2%Օr ١^Cu^a Q]mo\3Y{9 fzcru|\/V)E2Eu)G8+'pէSPZL,˫˻f#=iO2Nw.YvQNquJ{䚳ҥ+7Ũ^JJ?t .E b>+Vcg3̐XG]-/1]ܶ@ʟ]Q9PT+ BMBr$x0`N`c+G*!L1w}RBr$Y f"UY*8NC6Ű ^Y4/4|f Ac(A%-,\ KЅ%.e`4i@UUJCHrǴq,ף0hOgk 6\^xk끺O'NƏ>KN'g\Nuv;ozM .lfz+qߛDMD ݵl9BYLp)Ksnk9@<4Ll!|0QaeiKu}~CJ1r@* ,?mEc!j ˬIȖAYB(;<a >MĭD\4:9%wޭ|λܡ.fwê#蘒:d"ߛNuטִ~e;a =6e?􇭆65 [OjQM6#ך]3i˃̭a;j?7 ;f7ioԏ&àm`~K= 7aHB`l-i-&g!=քnt>=yMe/~ꩣVa>:wa݉O?lxy1iڑFmpgPtnki0^e\5K뫫{svsֺ{2ѠV9J1}嫿?>՟g.㟣7:-&/Wn{U;&:^~]wԏI\ 7cJ_~x0޳l9RZ߉uڼ5d Ubrg?&%6gzÍ0H{r^OW%黸%2"qCƱbq.BHyf00XnTZ"^F2se6tWikOѰ@XLF=/1|f񾙐3eL\}sSF p-'4sd cG1e2ܾ۠qEhGQc`g jGgj.((GNV0@JzO! \ɷ16ҔO+u{6$r]xCe 0NLI Yde#ƧܟL aOǪMeY$[SK1y%6eGA?{Sf{~w"0uejNK'}9)M8Hf^||T!JS ' #pm\Q(3!z%Nft6ɏ [z` `؈h!x|^J$C>"8+@"4!RuD BCJ|?0M^88JMB,-z*ho4]?=N"@ y:v4DB_0 5 \f~h >%ԏԚ`50rH-Ϟ[(P>BP()(p@ahT^!B*I.5B E= Knś,V]pp /f*Qq!yNB/\~pks#@-ڎFXqIq/}BƜ/t:{X]'W<ΐME j*CUa4Xqٜ] Fwu2^ Ƚ2AdBf?4f0=<+kP~aہV)Sq''x5כcǍ}f *ͣFzW0Yo4s0ʹlF;*+\ֶr6 )r) f\ξ@V8׬r瞿$Ԙy=ymN݇}T-~?U?h8l6ҠӦQd,>}"(i/G<1rSb* <"2*3S" )Tc|!"egD8+n._jGMvL M{r󶎚wOmZO7 /lQrRA\*`rSB̽b`` }k f[K TdĴA_A.9AHhb"X0/\D3O#it}SV*q <3тx>4F>![L UȈ&>@hţPWANA~A!諤Y(µUȎmkǟt]I3Ұ "|h٩=V9A1 "I(TiJi!d:$ GG#e9 b)Դȭ;=0o8aEo Ō_|U(kV宼 QF tx֗CNn ʌiǎ?=k{S6Mnj6ݺ# Dbq!C4Nqnv[vM޵5qK{ⲏ㧔j6JdH?m 1&dYޯ{zg"-uvꭡ?:d/72v~ pN kgrhI8ߜ]/w:whRIۮ'=jhΫ;tռֆ ):JPH-R/OMu -B^uŲ2ô~{Z#[hfͅBy!kSBGzFlBh d۲bIp yT.~),r4sR0$20ɈP`WJš3/tnR̖8>)ȭl=EK](iJ3-!,(|A"?`^+.`Jl)Kg|!d/ r$]`ѸF4Y'ؗ9c"lacOLd1HޥP,Im.[_(==Q_yiD9R{RӉݘG۵#%Pи&h,n鷀ݷJ-66 9@bٙ0Jh'{cUo 1XZ/aIRK_\5ud;%'l l] ~-Cn_FU)}71,f_}ן"юA=w=(DƤ: AxR|p:\#g?_AB"{a R uMl9-ߵ\s-ZihCT"#fYgBJ='ĸR1F~ 4PuGxř-B4kCO#i]Ô&!Hkxw7~ݏ#r皱/GN`3㴉ycۀ.OjUJN]|6;pd3 {%tw+.Kp?./ޜe}¿O2f,<40''a/N(wsrv1NN 0#~:dؓRh E^?րCPl'k77|Ӟ^h^l7퇞*д'lHp*5{[C[YQl")!"(vO~.(^;m2.ɔ ]/a@#(4%óFE6Y}e.PaE XdeMEV!8*Nj4-$rvܔ>w8[m"+Eϛ!#;#ד)yڀzzV"⶚2K-eElś;eG/WCo[uN=RO_.~du^Sv09TsW4|D'LazyN@ůtqS 40rYbiHA}21chG/ԻKU@oyCnS)iGJbde˻#u:T]yxŊ#CזimRݖ3p3CeEӧeO) iu88 `jD5:v/ם5*&d\g~ԷͼFtG{u <.ew{ LJXi-Ͻm,X3'ͼv0z s56{h4Հ m(Nk51+]B*!2JgeS F4Fast WĨ}q|c].={ ]z^е'xYIDkێl*`Б9) N,JI6ȌhsQʠ2^:d<3AENIz4>t/lucN]`b1ƥbŏ]/^@&cksjT/awˉk[F8L '!O 3׈4I#a+>#n)( s.NBxA9-MP5h9^ Re(q5Lϴ22&"C9gFjA 1RLD/WQ#*lyƤKا'k1 _ۣ&,gh!O#r{.:s*nl}ҼpN3c_/tΆzbErC.zVn'8k2iA>ZEډ(Z^ o7,nJZ&Pn)'._tJ50ɬ N\SqHKyhk'Z 0d`vb%5Jfȕd(ޔ03E$FD-EЀ.d1¯5^"eŒJ)]jIGG<>VhRIkj(n'%EQ'./PlXG Mk9u+H-7 j=QyEx* ATȬLYxp4(/WGCh8|7W>8[Nj$9FJ^+([C~wp0qFOr묜"β ,ʘ7Fl3vϘ;n5{ͪF˶'xkERjtRIeD!#/=Ģ'Nި!i̓Y (}jij ^kϱvnJJC ˋ-c?0{IU\xGdx>sQ{1U\2= '3b~ሆGY*ly4DΊzݞj&H+0nO[Wc22aq*iQ5Mi=Yd^d Q 8Njok>qC ]ᅃN'8~M-]z]~Z,_ܚ\*MI6'kI" e]QJDaA2wN;Ȯ̚$4 :e4$ ł  Sf׽s|:z1`Wb I@f3%@Lt% ϊ(%] ?L Q PI[rJNȟ z{.h!_狫y?{Y׮-5 *F)ڰEx,偈~>ưΈm%)3r~+u\h%b~u$,O6_˜M ]KsH+ ]E8;f=a["$%") "@@Тd@*3+2G>=hX<&l8 -nFu0d KYKe'loY f6۲Y'@q5J~}rLTo8͋;a]Q8@zy[@1cӥbS_ _Wfg`\E\'_#.z!3nէj#&Ccn]U?v2^rǽJ~K.1/>9,Rr6F#\);aFG UTԫ]ƌf+|4G(m0Ղ'a֑%{0%دW2Xtqy[Uf6P qB $һ2{!> »ÅQdOk6V7.k_~xbRW}u$Vq-*3*Կ~u>;e]z^` Lkڽwv9n{RֲVJ-Eڝ; mp'CJ*[wfn@^+Ph:U=M9_~ۇ\߾ߵhP~Vuh1KR8WJb"[it?^fCߥQeG^%ȴI័@uF+Z?gI܆,ې%q$nY}o 3K,(* !Pm<\#U9|ʽ l}^mb00Jlȓ|r@J6'ÏV%T"1Fb}U=_gƤDUnI^a0VjsTjP* Z Q"0i;fMN&,7{+ *54劷?p+>x_ab;9ǭ߮˱TfTSj8l0ϝ!`ȁ9G>`)4UiO"0cDJ'LbyI,1U7dsx+pOHdaD 2NbW HQ)<<O~݅b Wc҆֔O߯Z`NQ\zݯLŸ e~o=48aՉW9UpBBBj0D/7UJo`ga>dLK͞Z ^u~VC̛H ^dŜ% md$lˤ7[ *i'{GF!:gPUX(ÐpVYHL=BLE 0SNj(^!+XiEs:ǥ#Eg E'~?-ߌ"LgМ"CEΑn$)rUq (7WLE(.*RJ`.HhX ; ;"U%KQ x"I~`bh A]h:iK\PJ WC) /J\S9sY_!ch4V89A}YݩZc5ЗIL%sSEʭƘr6;橯n[hipԑY0EAQ xՄdDN6H M(x!sd&ʻOKzSޑ\K*@srr/?W]V ~ or_gɟ92"ӥ.#ЌPiRDgXxiܾFOnxpvp5=.?ꖛ.k1cNnc拯,qne2P,t?QAqkUwܽ ́ryG a=.zmsb7EA;m!uQ7jr=zq%8A'*/T _)l確~|gxmKޮb"w.} ~nsNMk?puϯ&ׁ(%<{9{yƄ);lSbbifl_~_3vKR#2∅W>+0'57l2f ". \a e4ˆ3ˉ`H0rOASMX8oMpS $'됬ax09ܘL[q^hX9{KHK8 pM̽ꪱV 5QҌ 0uI${?9JHpx|=V⌣]5ő-8vLoSZʍ&3 τ8qos :dmi0J (^P UFce!4@b.eկjTRt%o8DGQ:J1+tPvl0դJG$KQ`)D4"j~j fRf:_୼L 4r1u֞lf\y**![bP,{{{a =S?yBIb"IϚbGDVY2/s(R`rk98SP]!TB;O[mk^\8$a. -^ jyrkϾ/l(nrdefMu;GԽQ z%i±O څW$r XsgSa,&X1͏:֬.AM{,aV fF߿Mwyd[ qe[֝[PcOG+(-tlǁࠡif*u\5'Jiw:%[uNhi5+JSh;E4Պhe_66VdB$3y39TI$,aL9 b\ &%hm6Q iUdw.L+9/Q' F,PBSty~e!=,ƅ?OJ(ͪedV/?;5N}T{E=]srjU(j]HS\Լ;U;k<-c$umZMb¬ә_xsN1E`i/TrPQ`B:L ߭nVfi<(|b`2Ǥ˗IQV;0&ĐCfd/3.J9h7դHgT0Ňf;_*% jejt\(\։S# ^\<(Pz_fIKwqU8\(Jq%SYmq )&p:^t1$],S 2M4%2OD}ks396S$Ǚ#Z N*{W4 T6.e^/%6.[G#,~+ Rg)#7)oGslxS|5Gm< @Qsm4u-V_4pxׁFN!.4DAP"{ns󒯼[O ⢗r@%p[N1C{3n e/hL(wXk#dEK)h|R;Ұk sE:*~%pweGN2DފTѲS$N[l$I~Z]9afOE PSVh?EQ<Iv?8zBbnPR7B*ao$a&va擇K\2\E`_!b-p:&#lb߂lB/8wo>h% v㧄QX? Ҝ q%bZʪj1ѹESFwmmzQA>d7/9js-I O5IQCr(J1bKNWU]j"`]Zskrn62sݳ;g:"v._6*sCsVO7.5|Ea Be {ozsLnU.>sɏ'g{uc-|zq'E?[uN+ƨH!|Q'.. -IAP&#>` a)>sW-J]TEv2|r- Y{?f|KA_kϮpQ`Jq-]]߫=ߧLi݉ Gt?ސq-7C(nk=644:Edy9g ](aSVE񭝼Vυ1`jbԆceAC냳v10E!o(([rH.rG jЁ…J㍷*Azh9ImvGu~HJ'6|$BX>ML@S(#Y ?(OM6D5^JJF6RzM Zߠ_؀0YTPtK=k7p?đOGN XhT8M4Z6ˮ* J7UN;&qǓMvoN ]| м?Y}GzKgVxf|>kp,Ngw_6KO^(jl VX vȋTbŦa$72'(n04=2^uhDk}Έ\ ֵU[y[=ƎJ3ojusW ('6ݎ :xVPVrQ .U{/Fs=H 2ĆPѕ,c6%yp`N0}q @Ӻ%R.찣Sa Ң(щ08:R>IR}2C*hvߐ5NM'  6qRﰩ}+>ԇ v}0/{P(7}کmi[(FŎUSWW_U/ i$^R㬍s7#3ZNQmVd(N)Tx@]^3fDil_uB[AtbcyIYG"<"+c ?|Wl9"i.>]CwLeU+BQJ}?QyHBW{<#W2cH-tzBpc:r1z@ %- x>1d]0UP҂ 'QZ)f XG>vPsW J<NJ?BA5tw[?)W_޿T7 0F(~mi%yFHRsZ2#NĿp\{_ :%6a䂃bhӼ%@Aa~G&Muk*[뫲.oS87VTK>drU-U5Śki9^-F͡E2E> (a@a@,n2&wtA1\ Ѐґ2gXܡ'I࿔@=b0:Q9gILD4ZҘRPVȤZX\ IJ*MF hj=ki}8. NCEuú7^5N E;bM7U4$N \LȅLh븂! QW/TI ٰSݥh./]^0uhpa|gZG+PGPP+@4\ hJ@YᗌhiĕTMr<TQZdkkƐhD3_uBt+/$ãBe ʂ^P!#oi+Qy؞yi-8!kr#E(+dDR.F}J:\(HJErAH DewbtqtC A$U0YSJ{pzd5-8DE@H A^@4F "CwCPc*D'P $r- S =P\2zUʚW7nZ\{0 f(X/hSF,@("c hpY'#sS]9h\nӊS\L4s^IM*Ƒ1&'jROҡx:GPlQ#ͦCm!(M[>- b U] #=[:k#&_Vj*c{ X-ӻ -TPijxɭ꓎ Ѐ^Pph5s9tl 1M灠KF>g(X:H)KZ*$o އ]j2BL'$KԛY/0My? 7m>=[?~ss>%+ֳ/OJ`/ %0@Qzx2ט91?x ZL9)ڎ'/f03+ͅa/iS*42$5M&xSw88QJ*J;"Kz0:Y i:H6ȼE/ 2k<?!zol:r/\X@Kfj =*dqA~o0r{A7;:g`kϮ䀬^"`]-d@':3y) tzt<mo˨S_`y}L6Zu!цk">(&ʇKwRd uBO\“JXoӐMtyݾRZuK*rICae S;WxSg-\ݲ%tOvt12X_eقCTLh6ϟkY#^qlw=8ܧc7g֠r7{{}O਒9Z{휇y e0"Ƽ6Ι5wLNcLZcv6rlݳ;g:"%K~JTvg8D9eVmPͪ lnJ7wC54&p6rv?}|k'GՃÐpu;EptӲVhy+=7SZ@fǶGkvl k"\z#%¼q+:&pA(G#R8 WAOSjJ i+ Ӽe`]JT1 g֧x&)=QϾQ2ۥ^jH}~LT#զ"{U%S`<ِͤi̟.E&FT&\Ak,|3sG`QxkAmRtC?&7uܖ XK\-ǮWmƮV3z 6[;9pB PO8=ɂܪ(=ɃS.ѬqJJ9_KKJa;ZrR!1$.JpU "KmyC&_y-w<Ҷ<r1D2.(4ڡ9TH%Jr*Y/!o°M HL{0@u*iϨDZ A>4Qk|$*ܶB3s廒8]ѡ"PAx!抩^R"QC)MC <5mY)@"]H)J8{"pt7Al)VV}M2 Tji?\:{ JQ&7?YrC-f^%ͤВ|r$]mo#7+|;mp7쇻 .H69-{%y^r,m[f_+dl]|bU] ]56M|/?̮VM-*Q*\S1xMAtI5=^}э߯%~ϛ(F OҾ|r]< =_ra6oH /흻ݬ!K>5|zW IVf4zT흁;P*F8TS /(H1m<,zz3D(T}@nP1MeE+bʋ4u \6K7'ˏ.[fd1]~ ׷E9_r?i@2en&.c"иmKMi= 做m%hmcT.ve*γm8o("a МmZo]4u;G!7T׮vW0O/nGPqw`E e; )޶&3Z8e3j02vd*^O/~_n&3sڊ^g_ܖU} gܖ9mE ZI8fCK}r-\i'拮S{W7 g  vv4g ߿dWsw}eJ*Sq}O+n|h^7ui9hAThmx/6 ekB?\0LƓ}LZ(y|1!r>dg`*~ Ǿ P >[PLCa-W,1ipn>2h3 ʸSVy샳,T4b8[s6{⫽db0wABe)C tn$39= ЕU"Owo;g}ӣv~qFiVli25#F>1:[O>[B*rxӣX/F_^Z0= m$S}$O_4OAl3W;q )^۰(dnztmztm(Ѱ4ܛOn-.ӭT#,ŽW݋3KB J "gy~|IBWYH|z9s67z~ z6nf=f] +g+qKoVjG!7|EԕY{TszfĤevG?(D`F= zHxoc镖=b>|X_ DԭgtV :l412-P7#F[#&àͨ~*EX"@@1+-7 3$z#%FGmK:kYIĨ5Q%ÍuP*[4iSx:S.j^/F4AA ]EļT<0>L;I8\jPk$,~p 8'(AKozqL?[=hy1}.bԱ-HPPrs, 0uMԉ!AaPTTf=K.Pź9Z&+5;ܪ3l3`(!׫ܱwFTE@~嫾")Ur`Ir2g} ՘M 0LQ`HQzGȾ0U*or"CǬ.Y)׻&f#? ( H\)»d7+Z*3Od ܇.6M9=IiN& |Boޓ2ЄN~6F*O!tmН^eoiaX߼ +wus,zPp>{ =cwu7 '#l*ǿAqYI2*;(b?f+y8Rxa i%۸0g 7]^J KYjnLWNH3a:iS6fJk7q39邲p(oL $5Qpep mSqiLY)eHPZ΀M0s0eP_@,Lզt 3P:ym`dL[r%kѢЪy%wBk |ķ7Ps[jsd!Ȳ`ΠFH=XQ;C?}Mz^ sPF7#7"e_EsuY;N9=^O9뗋O YL4w\=5Nҽ%^iѢa7 8]Fko,7]=M.ʆzeww3f`Zd `M T04|5L#$L zt.*6Z=py*<֧~bک9HHK_jo=1nN{V; ׺t1JㄜZ撢)ѤdPJԺ PIe\D@f0- |%1ϑ˱9AckcժJ y3Eet\2dTI~.; ڷŬZMEӆ.O<_` P֋SaUFa յ]? :{/~?\pLJ>*Y8S f1zM gl,D:NA9D`'c2tC8#hI IDG.*ԓ$styUj^fؔ</9%yHSbu{a Q@K `ӖҒ*K[2D6i K|PF߭B-N#YGA[Mqpx'(ǿH!ivȹ-rn=QKKu)X *`ŌG`.XEOԔ9 ÐX %sF^ ҮM_nXvG L]3ςRgv*bn,͖ Hh%u}'ZtbQĒH@CriMԸ6U'-8 $pa@{i]`Lv{mHp\o>{=BkoLe+^pnwZ@N{Fq)ê_.V !l 'Lnܝb*CŴM&yܗ/' W+3 gLrP1P>@la^m춍yևژ׋#*73oZa F R9|=j1*f$LEܦ%  yL;Jq(?wI%Bvs-W2`T] T}7+j_oRhy Q8jFa") D/u*ǥ*pJPu+k ~1R2YNo#&Gp?+g?%0@8>k QJ>P?5%dtsJvV]צ\Ѫ^J((yvmXrݣ}X*Pg] |fnTcppGz/?"a95W~*I'*>"}Փ(k .E Odl&{ui嬤4`e N@Zr_pP@ w/c^?ϼN}p1MIAz \^g" ı~#2|Rp2,Wv8 ՟ST"jP4 Ef9#]9࣎ER;w N {I|qس*F-^R"x$01 &)JeBaڕtx] H@2kIQ]C(ͫH' 򡊒,syhD`*pyicKly~Q`ҳnƤ!'J$&)jnYKY+c+fh~lp1lAyay8I2Wy= Ue?]Wާ@jK7˧GWn1JQJ@GGWFqF :^w; L 9&#_^qTńlLF>TZ6$,[aؒaqf6D%B+Y5ܳ S_F/ZD⾎1իo@];/BrEOd5UmŬ#aO]!]ŜWc@90βSʝ|'%M8~Q擐ߙw KxJA[Icr^2|dIhi)E 'ItPڧ `GY%-t],*A cHgϴ[uP^!4|Ou# BM{ycPZ[6)M-|ƺkIq>&  SQL<r@wq7g h`}jQ& ,/)\1n҅H~xx[ ꏪB \vF[P)[Ey8LJO?G^=xa;@7Wu}U$Ыշ7. ɘW2aĴߚB` ^]rK("cA`O6 J2 RJPsXjL턙$13Ic91c0M,Xƍ4K5#Rq!HF:fKd}4Ze1V=U-cpLAK.&5Z+!50dP"w3n$7kØx)U&c5aN`GT8Zs` X% +bԜ:CIFc2Uejd!7J}8ÂO@#YCʕֱ+H*aHU=cS* iefd#C~ [pYc9gb"e;f@(pKk!1À( 5, j7It%ļD<Ё$ZL$:CHMW.%e!. wc$@0b̗b̸=ʪzCs&ZqY.Q9y֏7yJ ORoULH ͭN!n 'Hl'MxA2mNH s$^=eug#$Of£i8RnƤ!vI`|>Eա ccW%1^ѥd`v!|%k2[ggSzbbQ$ecuu.Z\{ht}G\B XZ:r,uQ(Q4CQ*LƒKHV[En_6w[n_7TI|Ԭ$56)2MmaFD%$2Nip@ 1jS{5L"6Փ1:Ts QPN["`ǰe:0a5q9T.)]8TY +챢2CwZykE jTci]j}o9BI9RKEx9zU#}쯄9ֶlƗBq?RS 2 SpF_miTL\n4ώڐ'e!b\R׬^п駻t B*]t408 -$\x+H\'Vpkd;%闭)njsJM;)5iotOc ]#{샍g_S/~l\ӿהiXK٫zG7orPBqͲ)&vly7$[.).@iahOJX7n96ژ$v1i1&ZY/?pHȗ?밻"ֻMף5VˌQMD/k@3p檑qpn_`^|泞ͮ?F6mWt$_9}A352vCNkiC}~d)04ȏ032+A/MmPvhUOe$LNd$VPTp,G_m0e89Cp34/-GdcqfAB>3bC>˟Wm}W\]%XXg +g/n4F̨#zFyhW9- %X ɸ`hߙiG 7W2_ޡRü+S **a'-7{+260kdYzgbNBi- jK'BqH˧nC?]$)a3u_W&iCdXT8O݈ž12 rHl9>X"G Vjgihs? gEaK~ =3b~^:wAE4Lڑ>♔7t<昪Di% ԈG{$d :_w xK>Oa-._'z.3qa Q͵q/j7;1oI<>y!Ϋq^}hΫ}F (jj<aN*t\Z 5C0mT)V-]L,Z ݁^~*/p#A+RfzѵxoHBɎ򛖻^T)3ڪK$,_`/SK*%#x>4ϒƏP. Vw}s"h 3bdl? ]/zo0}Qw>s)ꋝ hb"q#<75B[vx QTaEJɯB$ "VkT +N)a,1U(,Ÿ(f)|w?fZ04k09"嬔D|G[r k(QrX7kܦux}Y x Y'4BPb76H'VYm!14QaEK"a!kDh/) A(T2ϫ *<캶B."JkDϐ6Yhd5c4u}6YJuO*SеsaW!>^S.`֘p+F 4嶒q@ dvMۄF}Dp8^ͯ9]Zqj0b*{O\|sq@P{O]@vRΟvFx|d4{zpNR'L阮3k#ê([^%kxPCUŜQ>ࡖa5 Vt>P:9ž!i 0t?H(9da#t)h*~9DP/ﮒt!P -7:57@A~R!8\UfPeWT ;x \T(^B箲~sA>2 *wC &,P.ynnPC_f??fIȉ1t[3N)* =^0_Czت XK<]1exHEk)nhs61kgvƬ}HP>*bKRwmH_kE f0|:k,9a_-K(Cm['vYUdjc",~Ll)#Ck r|Hj ЛXjB߂` P&tDZcḠoSQ?kxLj7捺[pÒ/zt笱v_`6c.;rs9mid[`Z~sXkq8 Gcw0C*i.ӕYUR3*΢QZjKLZrL*y&uECLZ:OA1L([7y N w*Xpٗ"5ꗰtcz1j $9$;xq%7t"5ګlp1 6Ne \v *m8f[.7%LmԾe ܒ\b݃_Ʒ tٗ@w/ltyV!So(VG1sP}w=xaEXh'vݧ|jZ5\Ʃ>%B|{ ŞRGHrbuVň6)o+墭$\oSţhəM[4 T2覭ۄ5kH\铸ņt׌(f\R ;@ƶ S`2ȍP*)x< MBkb ,d+tStӟ'GD 3W_fϡ\qpf# ƌ'`ٖ9ąw^_C†bUY݃ό9aU-s#(޳=Nx\H gGLeAۙP[GYɷ;(pSY@zq[W.yAc=N rVd m6.ى1beM ;z(vEQB+7h:3ҺI)y\ ٢48_}sg嗏f݉m+[T M<%.^j:ҷď7ގ*[9_~IT?ePbR1D["AtI̳7{),),6 & :RfWb%PuHwTeK/U)XpOPGZFǥ^“"h !i ei9tA_.ӋHleYҦe(=&} p]4 kwԬޱi`R,'Nz1:ZnD]JVgd*a+`s=>Q{|hCvֽwZ!>zsYkLPu7`0M[7ܙja906v'XiziNH7Yݾh#Ͳm%9)$ZMϗYuE ڧ/5V$ڥ1b*5 PKxW}H0Hȱv4^3Ƙ!Wqv+S6}]L[w=.}2saF5_|JZ4?ebа֑`o!(Qpٻ|b>YdVjIwrGd7-V1pDᘵD@5;\"gq"Ӄ3p5]%B#&$ov [[?|[ N q\ڼc!y?ǜ[Pfd-w$KC,`Wc[ہaAT("ٴ^q,mw{u}0!݉:ٽF&uW7 =~]xN󴸢\=&ٵ;YӚt>bsC> ?)&n:==8\V0Iɽ3.ݸO<ٞ,0`ˌ;zHhZra8Lfº0:ZĂ=kVZƽkv 4}vwk7\-]/qߚ(Wp %ֳا]BaG/pþZF+/#m{1ޞġ!`=RK^o5n-(J)?ܝJmEw93!Eu>RHŬSXYtQ N95չ52J^mG €}Pv_suFc4G[֌1zF(&cET_Uu{=nwVm M;cg h=j١4Qpr3g ^!j$m,+ ݷkPkXe_Zs `ySYXo7WjxM_j%jIw6y@2;nYvیv꛷u! *艺{$8~,YXyb `d,yoBs{@6e[ӯyV&iYlVVWfoa%M/CۿR"545Wghu^<-ߧWކ:߱|ͿI?ʧVlo*h8Ip\<ڻ$)7|v1Șce Qy!g)0p>`P :J8ϧ9߱7x}}9/'mnR'>L'I g7$G ?f1'ϩ1il=Bl/#N4:bWfP\5%xEuO[5|Pr@4$0yX@pOhV^HF(:4n'M{'K)f∍ Q;izZwQxԚ5$hR`[ZsIFD6YHΡ Fk }\Gm-jwaj0W`+RIK)S vѦDg@A҂m JW;Kr"n cz6F vqf{n9;`;\RJ*)WWe QZzfXrYi"ʤ9J+u@' f.r"7JGh ,DRi/;`(HUNEA;dI瑩$Eh% 'crԜխ[-amB3 vE03ȍPҪ\yX`b@eu &Ͳ$#oS::-A ȯI Lh0Y:LE;En^)miaJ$\oSɐ1>'gJ$9!xTn9,q^e"y2h<31HXcK0%M`J KVpTpH16Ʉ\Y2loE0h"ء/pv:ûAR@=/#me##\:G 4c'$0وBN_BȤ- !rKH$$M FH?ǀ7W%xvMi`uL񣟦/0qӓxsQG`|ۏ 4PG=>g7}_~yϯ>bzџɯ|B?ݜ^|xGj5aW3gn:]T;~kWVƭQ?=9gđ"wfnX{V-R-yC.{9!rJ.XZ:.$3_\ a^VQFR,c-#,& j4iĮ Z{ΔLD-sl .|e1J+DւtvAѐ14.2 ﹶ\ $9NS&7\BF2qЧA(qPV[HI =Hq D %0<❆In0BhPݔ$7BI9'1cYhqd:sը>ɥ\-LgZX.|/FqK?%iX>GI&b X<<˜3%1RHXF ܦ4[Mx!`(Yj} M+ +97Mohamͷu!ˎhBJ}G 0^jCV9Xΰߗ=~uqEkS3*Y;uJRA`؇=%-gUz(5 Tc[@}cCj!MQABacgcGQ5pEjq 6qhv_Mz€cyAV-o󺌙AuyF_ٶ06\Z k{W \ﰇv6Z4Vaۧ'Ƿ<ΏT}y#km:wԿ}}Wm&rE/kK%ߗPH2%feZh'%QZ o,rJ[\Xb+8/<u yWVHҍzD&@33{miToM c  .X"-L?,P{"E<́{2&A3RclG¤%Sq[v!̐NRamɮ؉p`J۲y  I*]ez(q+"b?Fe\2xeeXZ 01 gg&U֓ƻެJ!u (eo@l#!:Xa9B;'%_$Li+BH!U(m@=,ܗ \I,j`}x:" EtCvS#h?L!r=brW2؂҅gSB%+ͧ7oT.J)JW0lp9Qe3+b c;uu~Z--E9(Mui(z|\2/A݁ JC7v3ɃD gFw3ajޞx) ksNG]S$26-!o~~Vp;4~ͧIB-gԌPhR/J0 f058`ʹ"~۫H=,/~(7  s33Ofqz?{Mdi'{;oOC,Zp"s]f"1Ri}kߤ/|Myu7i&Y@Sj@>joporEOS-כٗ(nEן5*kz(;$Pp=t?&~a!8(o6ﳫ!Pw3 w~pp[[P E~]O1!BS!KʳG@o⮄D R<%+ .1M<,9Y F iT*JI,Y7٫YN\f`i4@'^ 1<[ #@)?\I ߿htփ}^.>5HHB7F &ALף7rZ$_N/_nQzP{۽^: %%^>X^t4Du 48S"5ړ%Frh{q75rF`V'"LٔP79*VK zyڇ-1-!)ѽGNy@KIS"=]$}#5dg=zc2e lLϚ]Űd8קj >BVvc%`lX4x,On^]dCr\ ȿ<+Idmh$/5\,s)%s3roc&6eVqY7Z:8gX:^{/p$˜/&C)SXSD]ZDhəpUIqU$faIeb>${u9㒽sKcB0|X8CBJ'44؁ƀPTF3 $A%GXb#8UʳW8J^Nb 8=Za-e@s)As•J&!m9T7^:{uYb4ˆIvY)Š01)D>fǺ~'14g܀CY/'` #{1eJ/Q``0#! S f%yp Ue2' kf@j4ǵZ+/=is^z4}bL^~gZ1fȕ#lt,Bb8پ]$|c̐8h`Ooh􌱋CpDcuFHZ FHw kB*%H{oe5~O+y6M<7p(1j0톅IGBaM0YOa i"pm%st=zp [OFGe( ]#M3ZVfC3Hp5 g7kB»wf M Y &d.D#H. w:D]:ZZB bf{j٨e@9K<#@GKp^YJ {GIZ <`W#(:<+VTp‘^Рbo<RhE^qX[I)*oƱfRɵ@Ow' WdWRoMxUvb$M[,2:@IҲd@%p֛*)M9=-i#VFtrx+SY`թn3Gb`^oc%˳EKdwj]͖0#gWݪcz񱰳?j! *;4D5^ ](0]oDѩ=[DJmHVD+1FuMLr &`5a Nh 2%cǽbt^bOą֪ڛ7aoA, J+U,ۓGl9mJgˏ{.H"q콞7G:}`cd(X_ klof٧Y;'4N !(G%vڳsL1^H憑hi*5x2*,yc zK^4: ]+t 4Z4z#J\mޝlhv6~X&]Jz {܁@؟쳇&k鿛l 3w6v]şT}3nsàw1`|ӯ,ӏw+3} g3] _ncJ15؎QgZJ *Y,jsuհRv5lF㎒(QgZJ E dY.RIX]i}}QLFf)"foU.Qj̤G*Vi(јFbej4^l$;]#;P-J3)ZgZwPțobڒw{G,'Z=x 3>F- _N|v;1Hx&U-ُ9 wRY4+ƫY|cA+  'Jk5i/ɟ8cٛ0k :%*$sU pYz^u?&mW͔R64+Q;c0h/{epE!wГvT}Qܑ׸޹^*=J#ݹ99îԮ5/~\Ss dzmN_gοEHUɰ'HsAӍh~6)(s*vH@UA%QTX]TZ+*doQnRى,O3FRV' ғ b{Xq ЀX"c I_Ja:KM'jXv{}-eǚ6|Fv3!=[u8_~+tqe1C ڛ^;bhy2kh7V7s{ӥz|8czMp%d_s5r6 5?SD}!+5h~)ʜvg)ʜ4PQO+B̍,z18?=<f{!J?z\'B 8*d'Ae)% CZ!N:=հv3pt~;~ >Wqh r'+e0Eo p$R㻷 U +y0B<=Qo-P|DWÆruR\K)Sx\}8´g P"VA\v"   -ЀXEi?,; Җ}E'ϯស7%ћR8]A` /4Ek:3>r6|6J]YEn;^t1Bfa˛hy{ԣX Ҡr 1fȱITɇ}F4'?u! "{OD5]=(4 6>Krqkz'~_I%.d8NɂKj`ɂSDz[ - $6T &XqYkxèT NB& JQ:&c G*L%pČ99=Œq,=!fAyg[*זB19Ү0YXX$ϕ̡&k?L[RJ)JWƚB68B( a3-18Q1$GAP6ޠIpڠlP-LXPqZPk2@(EDV!rzh&l®59jq>| W aBO'8siڇ:nGpd5s:}r7%``uOcF.RveTcU~<6>Yu ޞ,"o}gWhO5B(L@>_``rLe[~إ(iFh(v\kaI h {#@fEvxÑ_ Ǵ_fiڕŦC8ϦZ>4Bgg&91< c^Ân;Z__mr@9ЎV06sfptΓ${W^fp;j\"4gDd%V <ՑΥ37{ Ƴ|/Ad\7D`(W@?fT*]RDWh>;@#Rcd_-u@ȗ)Z0cIܻ.ovXe uߪӦLW[}s"cbod?31"L¾x6lz}E_;(|M!.1nzpBxbӔn"Mtm]=ՒQ%CJz^yP%09 N!igJKW!ƙfrҒ`Tq_8'MDTr-_`,;P[ߗtDXu_RzPAѴ~i ,FJ6D{mlAuTS%>%}1OWEu  1t,_k ˞qZOhsFy"9LtcN*krW$HY>,#`-ѽ5 |#w5۫1NJONgh4q~avp&we,d*u-gcZ'tiwk7,n"kゾ_.ge\8k-}58P'n5wmmEUzHNթvǛuFbB IOcHIÛ!f`[@7r_Nptem(` rf$䕋hL1zC0-I}Gv[?Om^'R5!!\D)9D9Z!VV*Y3&`:,h_=8jFߘN&c{zI8\G?Y;e퉆'8cF7y9y1 W:Y?o t E!$hR R_t:XUIl}' *黕N_\M=f>˧;_shOoٰN m?!mLݹ5[bJIO <}տH=?Hkaõ%>* =3&@>^fR; ]ͭvp(S.iw<\T&O\T*&N2yjG+]zZ:FZHxs&:ըҴ0:GXX!!g'/bbϷh49Z]@AҮB t:FZ4ۜ:G\MB ARWqNړ4jMTi;- :hPMWXp~b*,֘<Ҽ+;\zwеjQz1)dDdLvICEFXjH KzA^x-3daEH10gxC@ KNحݚm#_ء#9p%fDr% C#{*Q V{.X ؿ12bDID}ڞχI6Zq]( ' QEQ$i c ID*T+ʈz@v5ՑD֒iV ϙ#M$(KX(madNTpy^6d9. )j9||X٪zO(f %q̒%M+ԜN{;bmBًZj")ݙj0H38!\8L $ARD\`%66UcLҺ O&˘J3k"x/Q=)Ϫmr` Vflp?{zru؇r^g%QuF£&XQX,~jdW [|.hU_꺞.8/xOd))tmnS 6_Lgɸ(~ol2MLG^d1BPƇ^De-rN0aRĸ%o`  ^sus6$0 +{~z3|Bmv}'葜 :Be+Fp01ŏ$G|$`}(S1LW5kz{pw J)o<|Jed,f<m'yMjH=@H*)jKHR.Y(!SK$ N#zTKDĻc ҵEݝRHw> `IFJ_˶g}ejA>HhR~^jD bƽU>њPqrjGzem>y'8z&o'|VN=LiY~ U|g{geKa>"q{Y~c5>; ʸ'+j}5WFY6 "•q GD"Z"=,$]o=7 Umq zx_J-TϸBJ>_}nq&)(νL!NQIщ˧;gn2JQVZso HּOk$Y;Y3;B嫗Ro4ݩ?Up8^oULXzقN|\ {ZED&Y7j]^:Ex-S[4!NJH+Q ˹nOM6TM+Uܒs8՜rPD\*%캓Ot8 q%;JޫAFy"c7Yd޾mw41B+sm. 0V,f-n=UaT݆$7~寖򡅆e2?2j˻Oһ'>ðuWIWTσ7>܆ K_c^=R7hT bD'uS;n5#[E4EnZ>hT bD'u/ڤ@t3+#[E`js7,XznIoGԫa͢ͱ5C-_m(B`}q][B[~]Z>f5{ @Gt7 J&3~7:Ĕ4į+(a86Aȯdp~A["RRou>_x^^mA~3&8ooe@Ah6P3,.φ'wfv2ۓիOB>MT ջ&/N|xnAk^7r\,u13'l,@fA9DkH*ɗ}=!&T±:HL4HwHi=7nyqٯ$YJCSn07lۣ|H Z ZnRAϦE:LfA5F]$# v7DD0W * ZeL9h>$d{q xɕA6hFVڍ9( XCaÚYUP 3^!Ur)yA UXeOl׆qbl=ZutUM78ƛfx#5_a2W'7DUJyσ|<y^l<ͤ(*C%. !S2ZzU`{& % 椀܃U/N~Xry:xt _Vm p ߭~3Xg_!8R\FuztbN(nj(1E9o|ά+ HZ! (!1p’Z:c ʻ[<~i"hU0|Yh)@ 09;^x2MDQ< -M4/YހHS#z['q$ZjFbc]MA`zX[8Hi(N+'*Qhl%N|O{[wQZ&rT9Ta.! I {1g 9}Id QU(R$[T+ (Is`ЌCҺÇ`xYFkpl%4Q-Dz.:`,6bXIm: j#=f Kp,٥HiT4ØnT (S_.~r!XiFB0$!uI5TXFіxU`( ́ŭB+9 ٱlI[z#- GPn[gF"qٜ,/DȔEɋcST8#kRj_߳aP99eN0>tR{noC,u|13!b$!mmXqh.O8|Xǿ9 VXcȻ eI$`ˑNᣟ9DDHV  x6_ȇd|};"@0=:,G8\ ٻ6r$W9K|$v`2r&c嗵l_%[-[b V:woxs z&d\i7GAq@u-v>FުǚV!̍uy8{`݄ `^nh"'d]//ӯ,*h܁00oPeQu9DZ#3i(#),JVC*u [Ir0r0;aI }LPM r3kKcuFPBlѷOER^$ۯQ&5^arcйD.3)Cf{X%VB _Ir/Zh =k͜Bmcqx{gfn"'W!R!5G[BxxJNJ,oQYaəF;z pz29 eEy0[S)4ޥ]θSYA!gXJ`!,RX"`0CM2oTX` AZQ *&;mPۧqT:;}%,jM*Dlʈ DVBv&0rjdZm"K \MQ%(BEUݐzi'8.Fᝥa={r!fȖ΃ &GvRi8ٔ2|`CA_ rƞ!Hץ[ң@lj3c]Wmo6-な͠TmloGU7&@š-*>oBhxzp$}N*.+M$6 nac1B >3Z[z;?@Z1BNp8oHS[ʐ9FOǬc|[B 7VO,oER jz Ԉ-cMf Rbq ٗ^;dWO>dWM0dPsjzwEҢm3FwEӽX-FKE0$bTsnڬ}[O?DyF*bwzp1/(q E<Ɗw1*9Q,-rVƈ"vٕ>SQo{(w|m^Dy1FLu· g˺QC˻S }nM_Z0W}-GmPI~a;QvM#Czw s(+:n"Tބ#PQNpH8&Lil p0wg_`އ߬m6J⊐fG.U+H峃%j>zs]Bm2hi2ipAHG`IEÒ勾$xa t/6(PW%rQ3鷉Lv'KOFr6]JzBƃEߢ@#<^u$WU_yU𴢩D,Imr:4b]IW(efjtkQ}o-9VZ22IE-ϣe f9<@e.AP({/'|mMW] \>$Z; ٗUCb]WүqHdAfඳc"%uC|wRKgM˧%tϓS,c-wR0^ygwP+?,ΉS_H0nʘFӌ):K h[SP)G\h-f#7jT: |:2ZIK8QJ>1#/p^)[1sѧzѓ@쓄&V&zc&B^$ZQm ‚$Wl>\j$z^y!͜ix7V?|p,6+3ޞ]_V31o{<Kv.Qgׄ*ֶ8]Q(m[b-p|6Bcz I$ϥR5=>ffŊSIJ.a|e iiȤ/^*ļ\p7(S\wN* BxJNJ,Tظ!rƕȜ2  #MNHaFK c\ȳmޢ@J;](پ9(q0vBpPIȩUtZ4R,z50Kb}^ҖIs/(ZLZQA8 >cL(&sέ"h&|.[--xl1c |.#"Lc#0ɕˑQRay.r0}-V! yμKO!y sAeܜF*&"”$sp.֎ ocsBԃXAt='70׻*{/f|d=$ [D~(`R;t|SpMoVKlɺg;F8,5B¾Tv7&؊b+z :4 EGt>ltnj.ͮ*?9sכm || }svLa-񌌠-Ќ{l6l|if.d]?#稜s4B-.rlh^*?8G=;g ivRD 5&Z*흷 z`RDܭU㽳E`cbk%޹o 97 w LHƤNQP#c}.ݷ%v̞=eNܽlKl‘G;ӧF/NA؛R4FK#ꐚyC TAywri!Nڋk_}W'19B8)01/\y hY<ђIꄶyf=I9M>@"U i)91TSϲ\;L3xFeb#}&ˀ|L,AgIIqA hKԊ܄ MLb5 0a&!w$J6ɹF9dU&g 1wJk S SgkeE"/nʼ?4>$19XN 0凮X?><:<oMBW{67`G6e f?K@̓{( X}24_}3d 1FR*.Uo/s+*eNRA sߛmYJu~w-4w3>\{=B8pN~H؜.i'$.~/ .7ˇ{b11 c6wO˺B˧'"jԘ [?>--2g+ށO>AwNSD kMlK+ls|ş Bo?.VjX}yrP|6'ɯ~/G*Lq=LK:CU[l-BUy}ұT pDOÍ)">E&SVm)@"p@<|lD%G 10cE˾WAd#o '=V=J5.1CYegc"Bkr~3JsZ$!S5ؓրdګROH{^&R8yFi(<?+VA˞2oX)*ڮxrH9BBrq1Wom6CQ:-h4ӨTjhwuمXfeDZfU˫I!n Ah8w>W4sO3@e L NHZ1[ØFzdH1sj ̨S5gǛɇۢGvЛI880'x)O\ x(ڜ<|sFz6ʓ?xl e G0-ܘ\08Ҍd6࢞qt,LY#uG ZLYjQx9UqHBIϋ/f΄Qx7V?僓z% {{vd/fWŲ{˶7U800u3M Oxep,3)Ln2 UN{@4˙> lak"{qltؕVV(K0V  cƷ/xq!ҕUNE(9/S[sOѩ7+2gVBkTKezT/5(Ա fr>c0f =cٓк~D7A9 hyDkM9mbSwSD ˂wzuO )CZ>W3|n}G?1>7d 4çG DargeCΈ!BOC0;N =`СKލ2ѧ qκavk#x; AasLQQZ}N=A>Ьu Xue,2Dg *(JAҖ#4tg!ywtMjv~~C~v}^|,-)NEMDC#)iP7ZSԽ_5* MD1*yݰ'>pMjh7N|AhԼt9%[T̗E, k{'.4NUI '&1ZBibZBk ,,<1fU6Ko|VBa GAiJp ރEQQU J!UVNNI [_7h .H@ҌZ͵:F6/ 5onX+lb [a-WՆ5Ij2*3O`ٯv[d.PigY1;-.r)YqR%U Iq : [ֶ~r8CUWd;]\s/]mOctLT)ҎJyO2Yں|`|W5dz .H։A;i-cf:hƈ>Dҁ`bKQ3.W v|V. ]@leeEIooݑx5@FI[roCfkˈ%XCCZX8i'w06Z]Ev heaNf%y@5{< =S| IV V{Cvdr _J CȎN_A+S`LEYUd *iCݖ]j;;zߌ?+u^^ofqSfRMb vma?t8Iٖ Vjz`l39vG)ǀO>_Qu+;h }+$ @NM =rh*}5S6bH▆ S!,h{zP sl| 1.{`2wcU \\\BBM˿(S“`2w8/)*s3~ C( {tv2^$#Zz!uQ#B>7#odB 9:I+k:~f&C.&GW$bdIٽږ/ Q+$ ~"jW} Fp-&C'?wh5=ea>k{C0` Gx uwR:څ֔g<8H ^cQ3xa\V 㯿hnWm"V~>Մ=:3Ȩ)uȧZ>` _~?=5`ʢ**t TP)IE31{9˷AFfsjyNY+8DL5% 1em~Pn][QYiy3Jk]y(DE'AO8|&nJ1 bС4AT1A|T27B:h]EjHIR8*|]&']CŔg*{nurIyѝ3`>rA =6-K74Ik<`YX0= ciYǠXs?#`9q1ܵZ5q u?K?:k[gy/6&AV6`o[dt׉f8Yhbh?8FgseǨB&7-_U\-dh0 xAL]{ㄖF|f Xa%JQ)Rtj޻nS6~k;+_ZbfrprwSk`v6 )~Z +n&/խ/m^M:n_Tgx}߾7bԨ_&zOUkQ",͋O ]Y?KgvYBuuҙ[ _JU* v+ b4N : X~J77WWٺl|Л R|Y\>7W3g_gß- Ȁ_1\"Ŝ( 3 F?ޠ $ :ь@rg4ߎAVNQ: 褔;oЁB뜥=VFe^V6rD0k*\Pe@ouFx36-o9fgin՞ݽ,>mz;(eAL .0?,<4wg^,ݫ{#b+j)}嬒 k5;>6h6V֠ॵ&} oTKA\k٩Q4م4ܾ:Mva5Y;>KwgywaQhEc@"FNH)v,z/a}k}ŘHw5. ]@\tE)cPl|aacش(k*A>iXJ!T:FKБ="fY]OÆO'iT+T(Q #!JVI&V|(ixFW7BU5?h;'t{7It,\V-PЌw7Ҭ,\]o/x LV)?}~|7Wqя"?[(ig7'jLT_3!,59<Ϸ`򒿟@[H-6}zg Ow`ۓ[qâ4QjL鑧LsX>ReЃ]oePIzzqgz%{6~X Pm<QtY[l hlfqeJ tsC-*V=/[lwME?z@$(t?09l A7 (b6Q6Aj6E$d=1t`~]*Q; !6") NU1lf#;Qx{ahWɁN%mڦv >rЪZ;`PcW_v;8nZŔizjC){}Z|e~{)(3$|gSl2)4IG Y;`gWRt?X><]$UTCAKe1D%]ЪQcF9T 7SXNT xgYIkRD+zb ZTPQ]zSZbs–Pw ;P:OBvKʪV$5C-ikc]dB8ٹg!K!f˾~e,K2MOv 6δ{,yʪf9F(%˙v5FBmk05傏{.G"w4-VscqƏ9?G{xX2|2':/>ޠ޼wzD/.whng6)F$Q9(KPi>0B#hLeR%AiC@"%KXw ,7 ڪlX?-6y{*}gjDN%V`үPǪlZVW5΀Xz~ROgW'j1P?̖=anm"D4Uct0&"&&$&bVipf1[R]DLD"fxs('CRLXV`H oOțu46yw A"QDPdmDq3nd$-vEI>dW 3ōǻaJ脧U^wA=7zɁ)>N!<Id=,vBKKsPh -Wg!!6Nݓ?_jg+:=k]Jm';̧64FȞhm2hIvZ<5R F TG} W/{Wܶ K_nsJ*ޭTݞ]v/-ؼPu߯HBR.G@3==Q6X8GlU]R+U"0Q \+'+OsvMUrBXlǰvr7 yZ=mݿ|`ѩ'B :X-DA]"`chAb ׂ2\hC=e޿۟ݧ(gWnd9&Y.+Ae ]_ծ"{V\7=[e^†Yg7v9$+42ExSɗ9ozB{g *`hϢb#QT(`pR QbScE֘lG#]ZːAljW!ų 3$S.?ebB59|3_OkfZi"p>ec MtjWpݣwh'|~s>v"u|=ȯT<'JĘyqxqg{kӘ!Rx}#=&3L c,#"Y֘.1ՖYQ? >um{ʏR"nM0gDEI_Ge=EhTr$<\Ryg d}I$'΄o|kkk*-]V@D3c JՕu92& 33Q,J'I9EA-C_>~<^Rp2?eS:Ufk->[3|XB%} *%^E^r1.}9]-??;Qb^hqbk4sH.q%WēkC RcīUQEЦNeVZBjtXX!ʰgaZ-Bb˺ǫz5 GƧC]G嚑1qcW8wWVܒ4Maӡ'А %’q-hghҠzLpɚyf-1!K g3,*+A6V&8ZAQFf-<#=o,jx&WIKLu `N\-f@8j ~jgS5-sZ Hf.MȤr.{nsU&by2*kRSbuB*O %<0ǒc,&0oFYbMeR,cL=5`V ,"$QsN,RJ1P 'Fk,%,M©]E?`"J)kd׊2`5Y/uM20Io뒻"O]Z1q'cx}?y-eeGˋ<=wۓdկP~R9dcj 4Y `\~6 ͕q:5}(kEI{2\:rE}-hTk\ޯwݑfyKCC1bNpgc?JyN.c*2.Ƈ--34;Pm ~Pp="ڰG! dﰾr 'UL Q4#c!c ; ɴ0ؤhLj"94~p.624GQNy(vpym3v·SD45=CHm 2 10B2g_yzVpaBxgV }#a$UYWe{~]72 J]ˋg chQʰO5J>f>caapibV2,IlHQZbI&T2 H?9͑(JBP$p`qIl$E f,f @m#um Kny j@o&ގdwUζ@KRQbNje~E^t1)"Ӫ W|z]>şwfV~v ;h] {rz,)|5[py( ^8Xnۥtkuֆ<,ZK41iY4ßO]yf Ni@~ȆWbB4悚T*wsaq*ɤ4wO#3\)IN$JO%uTU3©PnKVo}:{d40*E)HYfpV* '908O 8Q9UVN5(Uҟҏjkägtν`:΃ XnЮ"ӆw}s{|{~y$[]|m5%Y1Ks+;;k-n[3pT Mjii8 ”dž:av!YR#ۅdl2S lhxFi%p_24} :_;'gDL4"3;2*8Zo8cb9\+_̩$D3.* .7Pa%w]*[![@xw t5[zwQc:e[s0Flb_\'t;F ^}p.Go-p[ kù$cx3pT*.4)M)#-U!v)>2>I3T(^r%^WTCF>̋aL`ļF084.t^ όpZB1)VeR+Ct+pW0~|(YBfz LZI2ldmWF#3+`C8lڱy(fK- ࣿ|ލǏ~ SW) RP13r;O >:uL׉,]bt?~{YV̭ pvrc`Zq1*`ɜZ1CО9qVQV'֌ȣzfaX Ҩ,{ HKF:3"0n#˨6䴔V L]!b/Am0S7+m2=F B.%/W( B=C>! IKE_PDj$_IPi}uta<>!o 6;㻰|r;oO ۘp ۘp-&܊3H'$ :CyA i3*;筲0$B+GjCIA0 #BqEPiֆ:Wg` ;U&I&,r8DŽ֋\)Zx;M͏rW blL4cg0|˿كۺ*ͣ]]|2Ë~^̚VsQ :/**oL|/vtjHU˖r)Z%|FZL` /Q{҇@C :@,簔+YAMzrm*XbpTaR&Q) !b'Sr!94+(#+ ,rJYu ֱf1 Z`(?< .{ׄp2ODi="J"ocdM0X bTc^wi5Q[#+UpZ@g18kD[3Ё*:h0YF<&Ҹ@gKÍc-5č}6FZqBk.6ұ>!4<ˌ8#04gOwdzro](ICKkrgF+\G? GxE[@?_~~sp4Izu?oowf4a6V &|D> H۝`"9aF4Kq,yqDhn`Jy ֙ߎ(0qXA͘¤m8,*[xoը1h/S m.lUyM- O_nHM!ޞ#%V67 '{ 9: HCٚ$M yT1s͐:ŵߝwvca5=>gg?'S$-sV['yЖ@tTm\{b/i5*0j|Z]eO iGkTˤUZN(ʓڊ_$m*dKݻKdػ ɭ,(aZkCaև1>\xow1e$dm)i4QhN^R@A8h9p`[8׊h`^|}0=Osd >R, 鬽7D7 _)PQJ*+jY2Kل"^2&s@SXnVd^'BA#O~)g;t_VDDT$Lы(ze7WU! f$uhU]у2A1xjCtZ8XB_Zham/[ms$aUl#>jB[EɜKP(/誮(\!yÌ,e#1,ֳ~p1?tK7D?<Ɇ$yw.Y,-{s>$y@,)%b KYLc,)Bh3F+R+<xxi2HX;RP "MbهU`0r &}߁Y}7OGǵ~;qnȿTءr?ˋ _zqkQzY[t)i(%j>͙9Ms/O'{v,䕛hMiI[]nлbPtRƻ.uZݴwKK감Wn)6%nɪwDbPtRƻn7[aޭ y&bSv[ZnR1b:mlev-wB^nSc ( ^5pЇ_~<=#Z3֚PB7sul!c }F11]]uEyItG71k ws*Ck/nԅ5jׁRSh%E6*A^^ [|qxJiP{.U}H3Ka[ AXcTxVM4;{ W5WIE~WMX~dU&|Z8g׾M,;/.MzYca]lv YZi]svƱ:kݨK KWXfX_\&Bn+=-JZ՝8 +*EkV#qۀ*L}mI,D! +ܭXO:=EØ@y'.Q'7ü\|kh2xOɢnEFu[`vt 3mpZa"2/'^+E#fP#Ö)sc}\#WNjCDxfZ4̿qOn|gJIM[\ qWԅr{ ~0cg0|˿a`'oy_ь.>Eg/fM_Dۮ x1k{bY,+,[_&FXq9' —L9C 9j9)IQcL0mJ hq8Y<}-U7ŕAbNQYxZz*pk@I85wJso2K2x>#jՊ%VG * ކs !7Wg7Oݠgo0[ѥBd]pתip]#&IBLhOO]& |_a tՍ./.0:'Xÿ]*)E< E|"A?Hכ~{+;lϦԾa8,w0:ܟP2x@ !N^T^X  Fp(|D5p3X(ac='dLv‰54x$aCȊhs&05[fjD{'wD&¹e*0(NL4M/QiYuu\VKD(gj{Ƒ_1p"S,e37_.0$Jqnn0HufwKݔH=ͲTzT,V歙jwR5'åйĶ /tIqLB-XaAg%^#*(+\;ٺQn%o vԎL1e`I7 (wLt#yj[ncmqݺk蕢LfΓ ڮΉ/dŷ]jm\p1;MF_:e@HPh$v"U9I vr +0QV6ն%Xqj$Vs;4,%M ;pcW@!bvoz–h+ &E-%N !CsaX{ ^pkx֟uB,P׳4Vuӱ*B(r뷕/䇴U*.9bIA^2L$:SkjS/PfuY ˍFĀWiE 8+QQIY2 ySIRzRGB #{>6 p۪ _"1G&¨^6VL4T %n-Q]7l썡dҌՑto5EZ}i\́)g !j:d$Jq&jϯDM)"CttLvTGsyeMp5䐱naxMaE}z" sĖ ;\ {1cAzHO_fńCL( G =-}mB,$ HH9yW[)wѭ{73DcHN[JVqI{1"&M׳tiSu{n'1f;tGL3~"mk5K$`9`DQօN @Eݟzz}~ _ rQ3=Xa"sqQaMy8slOuK97*m1sA/8J yh ^pC.G8Sæy.w/?ZV Z>;999wk (`hTFHVi$U㱪4kJ% 0- 7, S`^EZ>|tTW{eJ(-w1ZܡҾ |yy)'?tKc*u]BJ+`m4F2lfH%DC 4ump}5.6GĐi.) n)Pҩv~).?kI,wBȣ> @1ExJKN%gǮp츩 g*阇e~mNfbԓPgfM[(R"L LhĎ,"Qj}R.,\׬EV8D@Ne.[Ux|-J64[W+;' _T e@IEU0R` E'v/P]Ep`pasWww9Hru˻P$o[eHYԾ{`a^|KՉj$pBTw_o֙;#8Z=%R nl'0W -՜ Yl!kD8a~G7bv| /k.LJn-mJ2Iw,q_2k%#4W)Ś"\3;#5Ρqh9R#-h9s8vT"NQ||(PУ9y{>V(%󎡙s;ВY 6N=-Nl%PnvJj;"ɆP)±FZw'2 -AӸOY,KJAL0QJAĝ{%K"*kdyk`Ԟƍ({A_l")Ow_ړsi?!k?W_uU3o~OfS(un 1 /5F+*Njn!ja}-][Z_+ YݸunoK2 0C;_{4wm.2rA燣zWN\*"r &jZHX5-۫g jy:!7Ќ B)ޯ`=7Athd*Aw}3_~(;Úۂ_%ĘO *6gb$>rs20w{b8O0P0^AV$9;x@%ɬ":{qЊ\-Tfn>Z3Byr̓Ѐ>JTd,\%N y@Uf́skfHwuɁC=n"o1/]6\@,Y W~rpeE~oC YJxC-R U `l夬xp.d}}4al~&ٟ JLFj$<x8Pp4Н!ķS:-1mut.CNJ̈lIao? Y꘾l( CZB)DimMKd8SRTi}I՘ҔD"R8l3Q*+QIÁNiِVآVWQ2U-(.1-9!\+m4UL"mtJN* 5VT2hMvDH۹ECKL-j1˦"ީ#Z ` O"@r+o^/E[,\5k|pwg !4Bfo~+ToW~ZZ߮K|&Qb:&v$J>4EC[9)MssHZ;@iRgG<,G:Mh~+nC ,7͍n k]\+T/.Ot}Ȫ㩱$ulIөR5Zm^Ic0C}=,Q)L%q̢eZKZ>]}e2C qJ2u4g,FPp낙W*{|<;w'J 61Aw]p>n} b~<gSj=|E_`vxa| 8|l2 "K2EZ9g>SB&CN*_Ŀ&f8"x Q^#l(Q©gr&<^ID4&c$UP2D0T Ihb)Ow4vJ5Xp FȂ& i,-khmfB]߇ bTV9x9̵ r̡wgt,'z~ %J{Ɨ&E )bۋ>ފT<㟠SӎTgS{ԏCEX^*f412R`OxKBsvRE^H6I;M`fq|CX֦6Z=/ s[<<"K 2j_8$Iy<J%qDaL;X"A wa;bX?"ٯŅ%sez.ڏ3q^.noyϢ2gZ{ܸ_Knbg#/ IړkG{E-^GZ#Zէ"xjfVI?/˓^$'g[)Kmj$p03*<9!MMj*3K<-b[mkGۇƂ!}ND9sRveE1r5XB'v`rDs&AÏvk M٩uzwڽMo_)1dzrs&(( ),(u^SE 5߼h_d،=܏1{z'=+ $d 1%if%[@qc8fkt-łkN>cVG_#$G+s&Tё\e(3w{] FZœsei-u.KCU/Ϥ*y4iWNW@e:(TExLGGF$u(k CYNSʹjU_JKrZH #!L LL<9ƾo"`BNӐú "kF\38 _8|܅V)9XՑpGL qFGK7\cfĹ!Z\Z hM$a`hlb1D ՀF bEV^ô<+[AiFDZceg_" 뽙ŷwrQЮ|eJGNVi;9GM``0uٖݾAuk,7")Ž0`:D?ӒMO8Jvv>nKºCR[M *XChkhXY 6;ǣmQ%x=\h~^\@sյF*ܺwn=9hWyJa(oOS}}B@Ғm_^ǽklS(Ib--irzbo9=dT/:.?Oir>>{mfW4CR=: ap:^%t:TzN0LiJ~ȟ:>E-`#r߿3CUorƜH*7+[4G&\ =3֘j=Sn\HjA81*/eKQH"\9 .`|Hሔ$Aɀ97FFRg1gsM= \`%'ƨ QJrj= #ajJb|ܐ&*pc5;`a~!J;V79זKPZ1KdP~D I5[!Fw^' .dճisճ)w%$ Z>qB€sy&MK%Z sB@Dž9| q*GN 9WGFFE)AצyM_ 1!GF}N$'0R/.ruݽf~.9j$(L:.釾">k DgA3DLXo4"Z_es!:O*s@9p|J*|s9&$jhHb}_f\9w:!Y±K{WZIuל@i-a,Z5~_Ǽz]pa] u `/e:~)Yo]/%"g6EMJ1)frU~6W.~zwvmL|9PKǗv=^()E- 5Nfd=ф툻J`]s2ʨRXǙx' uտ*" LMy XMGD\;9!~7(:j"Ъe+*F$T`%iW6`0YQ!@I}ɴ7^ +l`uw0GW2cNU"14ZDtf<J #ڴ| ?*p sc)bBlP8K9AGW1AA(OgY&Bk6&fc1h4bYN\9g)˿Dx ]%0 RWJJv\xkNg1@²{{d>YE%#CFp &%ZWUf  #܈p:7FтUʹ; AA#P Nc꘳*hM:Apr8)o6D`>7`g/gr=)as,"B`S(DrPBNؗD"|o.9^h%HYXHf)vnZ2߈wxƀ!Wᕅ$ZuuYI2Ǜkr|~"%]V<]ҧG#2) x 4 ʀ35 bg{jfjQvl0cw ' 8&( Hv_G& WO.oZDheɿ؇rU;f+b>l]9W\h|nYظ۳oHܳCc|u9/;ws Wcw0vWq[{E1QGF/(/Y9ՔmX9+Iϼ$k Q!sjE"?c8휓Po<'W }OROhF6zc7Iv|k&pQ" .0sLJ1w0ZDxґY+$2G1\5I'jx!iɗP. 7ބ&TZv'AtքX~TC4aC~+]1H^n,4<ߎQ$ }h컢i?O7n^;/1aVRo/{_4}ƺr*i+pp\*Ugy^V;w16pOi!̭'{% 6%$=ҥ6V1y ѾyK]Ha4)fp73r7 &`,ptu~ros?F߾~NBqwUAJ\\RF:uPS]YC $)nHC(*mݲ&zX7pE|oq'=T0FDorObݢNcjRzx-wQ)nbc]Nw/3T_Gxx47Ej1љ%܆n -*.UWAIRVEç`||3ۭc, rY!IRʅZKmJEP)7-3aU?ln>$gNHCQ]18l=D}tzg/`u]+_88T)Q Vgo?~ /l"`-ev<֚2Y!nDڎs'RPy"5a B:nGɜQW񯘀BbLFӛܯ~̬|sUˡ5Ck$Iyh-?^g BJs#& „k sSB磟)B\^Llgwz_QĄ+{W~1Z⯨QD~PwJC~П&?r)XF]'l#uTHhv*X^TޕY%He"~v[Z yў !kgD:ʘ1pk0Wcr" ]̦ɋ׃R5z2c}V\uȹ\+aa0\ VDݳA0Dl5ýi+1`ݑ :}Hc˞䒝m\R %NŽRѬgZb3u^ D;ޑ&sͯv^ RwwiҫnRC?յΡWC[`% K[-pn8jP2TUpϥp }a|v;{]j]m6+r`)-=ܷEn,Wٖ,J$%yFl騥SEzHi #aj*N)'~PKJ(Q\[Lhf*+2ȕ 9$2\a`#,4*^1^J8!t'䜳,Z((OQD{ƋHS"@*ČUZ^!l GgP7d^"l5j U&IE ef5rD3`r fS9֤0x1Dv%_9vZ4.ֹxwB1EdchG͘} x rȃzK)g!:UCA:r`8 rHfJqڑC/E@F(l+eBSN\% T&IMN,gHQ% Bgљ )fg\j /.~9.(b](awe~zZY /'uΠQ=,N|9PsI4+y2]d2 9U)Xf<-( 25o <˴91ER:EZXQJ_Rz5UFۻVhQG\h`7k4/ P%+^j"bbfDV;SJLS:)&M8"n|}ӟb\  K̠ lXZ QRև!Ԛm+i!|XiY #I^?j6P־ ]h;o^.2._ϯ?ϫw3wNpNU?8\ٸ;ˏ`fv1oE?h G Y&*)=_,R[&QGǃZǃ1;nG+YBFdAW%kЮ5 6PSR nALQԠ>$$|mIQ_17N*fBΎ=ptV&v6<<ꓑ OL} zN~ ӼX'gOn/-_a RXY̠OxغKnNĘOZ27RueZ,$[P5Ȯ5B(rL(!<L'#AL,#FcpՔu!.s39@ҟhbS>eV/f.<I"Y|;vJ/xmFy.GmrYUK{ѣIj%x٩l!iIMrNT:S/k/aQ=dGd/4FgB"勗Ig 1YR$+$E5*˳,> R*SzY0>4XyܤN:$v椛.f.]3@o6C#^,?o9fwzki|q\REI!`4*H@DP !bTr ;_QaIsK#!ʍ#Xk]B%c4@LC@i )dk. 5WnuZ P-1=4M(G%.9j_34C'. Fk,2}wB87H+|&TV>3FA tK{jX\c9 6{ؠ[NfW81vOQgNS{t?k Fwӆ_1" #~yCA{w?@N>CNAR$z :>kk4ښ9=z+F^FSQtoPdiӠ8f[} ;NeZ^~w'K:$3{t>C%mCXͶ(tu׶HRKߎ0 \kME3&eRHUnV(jj,h"-q O Nr\ˀ;5߄5Ga;yE&38M^DHhv /.;_nlwÌ(r[R2968D4O$3k7G,STS듞E>4i(iڐI*oжOj1(b2k}otU zMCQU 9-1g[堷@p+'N85%y#o7OAL3U(LRh%cؼ<7aP@Xq R="nзml?>HzH-gAΊbAGvW:Pݍ?iF6<2kuV~w?>$Mv=Yq q~ 5~H9ҾĢ,hgZQ! pc1TbUT\ g*stBA7eV͌arSƱ,h@0ldh6PLNI"7!gzLWτ,'J %OWy55>A|WϯJ} u[kn~rRY6,ulã~eQI,oz9wϗ}]Kddlt;b E t~Yό%$2 1HH\Is]u)U75/Q6<&kE xGA`'Bp7ZLD4<WzY+%O P4pdic#KPp IS7nI%nX7Hz$" L+kwhK/'+;ZE7[P)c*}I כe^ 4<2dW@[Uc,$j䄕n>q\M[:ER&gk smKtv8pOhe9M{-*z*9SL2g e*hjsq}CKw?xx|Z?~ (V>20SY6(ړr y xd O1FhktlCrS@_NBNLb)U iX=wfHFT t*QEVz!dǐmF^kڬ(r$IArT"y&Lr ڏv͒=Ý=ؿ>o˳53ƃjZ'=Vdx"DNGuNZyT7Jx*`<;!\zTӭշ*c8P;>9%Myd>h̀/ƈ(vܻUJqv7j)cƺ}|p;^n uo/_Vyā>r<_3U^SRyeޝz<8mC \L  p4;3ԩVT\[0ΟBBA#*^v/i&သVs8< |dG+`>4\ b>Q\m201Uӹ]1^{G/b~$3LLKD,S*E0UL UBj as2:_h,=@͌3(O(.;~nz][۶+*l= -{9/IMhD#)ǛQwR )j-5I˗_=su?ou(1a.5c$cpèʐ Tx/K+dB1GpšFJO-pjg4_zGP f^[L(83Il;sLF`o R{)j9r$s4oR>qxK"'SIҔKPg4k-SYR)`*3,FD$ADPMGtÆSv/n+g91$V5EʹчG$@r^qE 0e- \J BAAN&؆_FOfW BƩv48@{L:nދ܇ ϭ?x# ÀO[ZcOsvAI>P肙\֘fuxLpgs=n1l!a÷(wX9]_5hLiz㳩"BEbPHt"QJϡλծbZXU !߸FZ,i<? W]ƒDkۡ@ DC;[lZP3jB"nfUx|jq6o~JuLU''˂Vu[UE7i{\`Ņ"0mǝd1=N8zxB aY]ԥ=nWwOfr;'G߆٧3F+͘ c^?P1LGk.ɞ֞營Z #Ibxddanf t$BZyAYݿn\`?>l"!ΫA`jΚ} 3agqou^( p ?|sr++dnT6V2TĿ"̸@5jt Z?WX}c5j'6IʢCΧf>ӨYIr׳bcUDAYoS8>RA3f\tA 4ll{:t UP=|O`!8C^P%Iϧ1Bir:lREn:ceP(,T˛*v-Ċõ/YqRX!V(}|^Brc[ aʱ[0nJ9/p`IMuRo1UuՅ0C6}UfG?ӊuӞg\*{o=օXSepF t[~brVX' `Zz^ e}feR(/n_Ko2`--k,%ipkMpnP*jm~ݞ8;&ܯ{qWu~x :A.aV7[ )A-k> Z/w*:Qu98'#~ -Eպ<3+\~.k3} Bx]_C v\k3֚uLWmWRh^$U sՕS,ZE018QpWv.<5^1ũƧXq7 Qhbq Imdy#E^HÙު\2$A/TIyJa21,I9쵕%EbD PKN-% S#5=PZR]Rj CW+c,_䁡 %=P'ӮYx3yK7Y>cПyG=&NK"`s a‰fDJ]wW Lep  a4 V+Pjٚ%УHN8#kOߺnhN˒rXFЌU.2!M4K˙cXa+S0btý9Ԧ#cA$U֦$ FE|"j" mǏ![h<&&XT ֛$cQYkyuMI}gpiF@Rϻ -^d ԧ} "zV%#y˫."KF` )e:eGY;%V?χC>6`i۝c FJ.Hw᛼)|wT'tE@TV\XCUيa_U6Ypr$ʼnOSY)uA]~B'G:"k9V jC #1!$gMbr 1T3[6rT^rYe~U|Q4ئKn:@<ćܤe"4I74Dx۵rwވó"w/B+\;Wk`j0e<=OFrbgMZOS,iRٸųFSy dYõ8^D;#Օ~UZ U|jc|ԩ~䉨k\8R[ݙk~:WV{;qJZîx[\jZ|jeoU9`D9N"5aCSˮAnH9>Gv@~N)EQ\3B*IL*yX*|% ԾĤhROvɲ;8Jw0}ݎHF\8SR\g4 [k'2\.yjE>|jՌIEui5&P;טڏ("{c"_Yb kyJJ~(dz3Z ƍN\1OK/Bwk jJt!T7W[ *K>vߒOɠGPx ݛ_}^Q*o3?Y<v%#.Mp}dtfW@]gq_`WpT Nʡ[v>G&B3Ij6J@0TxaRx` ݂̾HuHiK_pw㮛jfIͩ Q땺3IKÙO*0 ͤc_7\0SCcgg<g-&_|`6~f0] :3;_9^C^~x%O͌(%#>d,r+x1 lx;=٧AwOCbʒa *ݳdD% 4%[wyI~: nᅔ8xƳyR "D;+&!%vvvF]߹ܻse ˕N5§sQZbpQpE!w9VS˭? ?gڪKK-k TuA^p%Qao^ Li%颷BXb?@t])8.jmla&)dI)J\1f DQ* MŁr zFD)D{KdRF<ð9$ 띐*+$0 fZp5 RlqʑJ;WV{#Ȥ:6E[D;:\b%rMBiTbL(|{LhL"s@S;rĊ]#VR[rϋ!' μBH{l#\Bx3w [_ZQL[($[BӉuA<Z k{>/T%7W7*7azR2OS*c'kt:CGyFYy⬳sn 8:!'TQˑ&:SwEstFJr21,I9쵕%EbD%N+Ev;<5$[EvBBqݐ ($vCdb+EÞzL<7p!4$k~}̹7Y6^l_8?ab@iY)[@G?>UBΫ߁1i0`{Yop.VYrs<(IiG'2EH4HVs nbG/y)g;_%@3QH96'bMӁ˸ {UΛyO杷,Q1{Qtoxe d{x hJ{T^#f _dQ`*QA^3HƿxF1v1͂ъ3ǰVjqڀ1pVjMPΤ1oahM$({no'm6~T64S(FI3|WE~e_Y_ض,;fXt'l3[}-ۤT'X~YkUD4~q49=v & X]:Ĉ { %bC)τ ڒ1X+ZLjW?]6'r 62sl8M2cr-19>tpb4BXK[G$Sͱ4(,gy~Xd=,os|h'qRhͿ+ ^Z=Oۣ&5MEPHoxۦE3vj3%+4]k5T:Ӣ'.<׸x!cM$p:杕ls52Kse`^1`NZI0[  bPE4uXA<o.,/0LH!Jdw3!95eܴX!:c(*YXA q-@R-wTEfp =Je(l)<@t#Lt4\ǴG{dGSxBBI͒siz謰fTo H-R?gPӠ41H)͹TʮK6Yk~NM;4%!d}`_cbxL?"w9bGWLƄq#zh x:uzŸnJze!WҒN"R&+bH cGAc3_瞙,0! ey&v ~&B &>:iYJI})-9>m'>98Y'dk2&hHq4ɨE7.Xg" +v{" 2%⦄h]z,,_& yP4귑cbW5|X@8c @Њ{?kc3lpVi0J$YOvBTSW?ǽаytտz1 >Ըg}1jZNbA̦,&Wpm\{`(XĪŭe"V1*$X*Aɜ۹<{P)R_*{c,/EYmp_?q~WhRb~X:XA]G1bt]rkN)L@S+D`9[Tr00CTR+ (pd&8;B_,n fvɏE C =0ˆh iX k>ҠV'3z2a9cHH:2F@t]+ !EY] ;B$rTLeZ8 6X)cOuV2Q-8ʚz gm8*Db]/(бxߡbTVR?)JwuE$*S K#XZBQI4)9NM+#(V;S%RVS+`ޗ/%2S1v1b1vKD*XR<ڃڧ{mXP&߇bg*cJZZL-x * $s@ \ׄ&@DcE&jў#wU6RP$&9U2cS)%qv~)aMX=RJr&@'.= 4rLZ)Y%IJvbJdn-I,K+.ԩƤ)1ΰӎ4럀=ct@Bhx6>**Cř^x!8x4Ny8CAFG%j zP F`u ^ '$N}i ˾4AWrRBYaqIRtzeçvmzgު#w}oVzvm&1ye|0"ujnoWDy`ϑ~tQ:"AiCFc*VPurhg'/>1uDHKa=)S'ΧG/Gd( CPl++/)krdF rGUGY@7tD4̓*^s|TOߜK'{`7+j8Xk\1 ykL{vZiNhs]&@~3^/ =ax'S9rgC PZ7˃'5 `]5:qр=UĆLaBK~S2=E|y$\mw7}>_}ھf P| =(r;A)Q i dD4u Q\G r]  Ik4&V\Q- Y$9 kJc܏VG9S6|N7:BہZPh;xe}$\!t~җa ~zw)__ݯ11VO}u ,$ srqov39Vh) {kFI9Xd5.WD2 Z!xP;ou_#S6LUDܕF4ba%TyäDa.?HZ=HKEyÀ&"9 qB3ϥJ,_O"M(x'愲Y| T"uP$;?yG{z(J(\y!$aUFf}cJI CPx1m_RK!R˯&Lsꔜ`&34ˀBR͈٨UA腭Uf@l+mk"z=e%rF㘌$Eãſ06 2} #p~5B,xS>B|n^D8m=]L7kU]sktnΌ4揪mնԶ~r`1lbխ&7]?=џܴFeͻ][9 a ƙccgȜ2"l C9猩7n2 nBdxqc??L|]2 ȶDzƜQIZ0Rb0Fde ֜@"}DfB3lOVjZ u z4ATz8_lhZnG(930mht97 +TEGU mPqצK9+նj C/UNc4? aY&!KywC2NvTvcD=U܂UmLL}UMwȵm䌾}ӛ}:׸gzva-P 5L:YZ&XZ5 GGj&/{:-tDQ`S))P~ D%Ĩk\,C!J̭dg8+̨3̼*a>sռ`)CM _zW~GZ)-Eu$А> FhA"~R??Y.~h6"zaŕ<v7zSry򒒑9/n_ cv7pwLJ.z_Rin_?FGx O^ u>}x@W=' hg!PNd*ds{J߼/r J0j9G'[XK>z-,C3\~^~y%FFzP2 k߆>0cP7^><Jײtv=}~j}iO?z8vGN0eYtY <;bJp:jge49 hKREEma9f5ؽ6&I׾yh OKGVZz|+ ri@5& ;$ ت}2jQAҨaZLE 4rN$ò=d H7aBqv΋߳k3yk+Fe21)':+xs?mPa6qΫq6c3o#*E-wYEdGNaCA{.5RNV\啢9pz1KbԶ,vEK bUP)jKvd>#Gtv,D?NY+w.P\hR6R}I**Uʞ~g.Ve;sߩ(Be;qNm@YfZT?hw㡗Z͘Ro[ZE `ZXSƼr4fe,t'+ѓFà>w )5s5h -N5Ў*9{Ë cLUEJhN:Fl2݇9A~!/s|}HSHD4mFg6W"]1j~?M[mFĺ;{b'Wa1'NV5'^w_0z=:"0z)TƶzJ{P ~1sҨҬhjTL E5rz:* MHB6n_ :=eޣ78 @ |iK9:"):]e$en}ZiuA?Px?']aQZ;ƖGko>УI_ibU>M 9i#/@6ͧuz1N?-eTT1K ax3slMRxccp56k}If_EBڷweB|}Kr!&F)%YR4\{{PѺZe4:1/UmtotWQrC{=&|^/7]a] 0g(}IzJWNJy\j)֮1 6(F9xIupDV -b>sXhlRc*^:o=0ܷ Im~;2 ~Vc,*BqQ(D4ThNc&em oΤ"ke5R$ k6?FQa77[?.~)XHΤ:{w]CU^Xn,\/P|}zJb!zz^fqԅ+Jg;H XR 1.KRka ":tFn18B0ru8}"ҖW~9HA䐖aoU/ŧx/D~쒇c3!ڋd`=, bc,h&PQeӊ'mДleҔL;W hVhV&UݥA3:B{J}D)o $yO`%?nyO@.٩B( vCr%)k.UF `{L$憁QL(QKRe#BVt)X{DOTrAѾBˁL_$f?wKi0jSQFF_>" 7Їۻg>f.2r4eo]n^BK :[.ya[fqfx琿=Gl>}B # Yw)ڐ>r3 ,bfGRyv/8ڱ]\JDWguӑEb*n:!o3Ji- J[bUUt .9|9:ƳW((!;Jͅ<;CSQ{8<ǒF)]Ff-RQLt$$HFDДBKG8JJ>*&n,S:AZfܕWJi3mW`ok"7M>rqNdl_e#XU2 Eя9ڦL{y'3vUҧbG\i-.oGej1H3gO_.kpL- rRpd]MH^/ 6u+HX:A`saBN|r%ll%vK$6f庱n4d uC3r;r-Fh-UJB.IW[=?fOn IV=|lOC>bt;MzgyZ<,P;goZ`)2/u8doy lIQQwKChdz'}6C Hq9l ye|7[Jg5gpgWc˭7ݱr/VgqrϽ7ٻ畵:\5C,hq5-O]zN;r4tNQ.izwiDz6|ԂbH,>]6CnƗzẋ,Av~T`I "xiW^˘7C9*AL9gr\$!?]H /u`RQ|BU:{W xj1V)_[[߯W#}k!tvatLӷJ}) _O{ߩZQ 3v۠F'Vkʫj\[0VS'骎$e%Ɔ?$ToӼ*yULU4)hcNZ. HB̢iX Ȕur! TB:)R~EJߏwf-bּ]5RjR[8+/jq: ix;DW)8l5rV:p@Fi<¸$Im h95|]]3ҕQ1|1,b\{,KX⃅(7.8`K"e V x"[ 7siM<{ll07{w2f>p)s5e/Vyp\>6'"D0{F Q0ce̢ZUfԒ~{VAF@J'{̙aLLFLqE![?Z ރ/Uq#bh, ,pܧ|'&؎}~dkiKj [n8UE^EIή +rY(WY .73դyB [\nDlUr<:fqB. w 6Z0& ͓I&Ws9K;Zm̽mi|^}g\ݚ&^"I>.>V; ,Lx}OiM-a棷h%Z!ZF^[꓆iƙ&I~6I(\ɤټnބjM}%˞s9쯖yY?ԔtgЋ+'(`ƤrqRt27>Ҥ-!1}/G4;G9VNuf_\]G^|}gvx3;i. X/O̎ëgI&$Y~).arDe8ǓJ kn`PŬ G!4Gg5(ʩi HN5cEW.Za^X=PNœ(w⨙狷+]L^)"}Y];g32la6-Tjlze3XGGgC^6 'M *7OKY&Oaaٷ]_Kô>d B)h?$v1d ;]QnBH&,ߍWJ&U4nB+9\w,  j5o2ب0P9lkX ̩_C7`@3"{Ō= )B޷"p=Q=#MhLlW'LPQ=ٞN& ̛*)󿟗QdeN#Mrz7'en$LBxʔrx@J:R40)H,H qsds"3iaa< n¤x  :_h R$9PIѡsm|=! FŅJ$0n"`؛$䧦f)C B4єÙ ()-l٘=NiMw$! H P(͝vǣ difc nR$ v`*L:(^cZS|̥ƪr LVis1є_O(S01cB +0^ DyEjW#&C Ȱv2|ܝΜK>P"uWRcٌ/9y"W쒚ZEHǶO9w;WxjEֹzZhO&坫ۧgG0"86Fvʼn:WOp&\ p$NE.sA BYɄ/wUP8=IzY>ў*=uVc=uKF8S}O j<MZT9Z >H S Uca- H۠5Zƨ! hDQAjwD 0j!Aܢ#s\2 t"ʁ4(=&3i K5/:_I]6_z˩|֟viEYIw >v+=;;`k:NuA9 \L9x0$3̓,(Q1?s$=]yzY(iSc!6ɡh mllI΍!SЁv:%C?ڝT Θ:EIL+><ί TJ<NuTR:89lnt0PH1,>C$!v5ɪӆK VEE,\!i%# b^0P p6N1h I7Dz۶TPgd>Y!iI:gfW%tG]_wK ~=/97nZWnƐ;NM!E״Řvף?Мh'BO8)Z :.9 +/.:F#P 6H.0u.slq)aW *3%:fiјFi& 娉`DPyum5& MVxڀ76{DB:E2;^<|5GJj}L%2K k^7*] o>h{z|iJ| b!+Vk'*,M͞8)ag3DJhA5B_O8R k9l%ABWH5WK*{9#\ p)gA( 00Fstb Ԝt|2 @MBu_ C |jN7jOώa2 3ϿjCq5 +}%UW2s5P@/{Nu&(P3aLL"gH&.@thXa"L݋ݶcD_jǸa>nP}*|2՜29DPT 䙏TKXijrMAOJ5:Tt.ɰ='A[ebۚ(E>NЋhoXxuV$k[݇Kv~[*hvYƾyp G]M~~n 1GŅj%r' &2 Єa"\Gկ/W&ٸס z[sӼ+"q>wwEa` •iY+gwjqiSV3ߥ}hXw|o}cD'KѳF1fEbuJjh_E/`W-uͪ&Ӌ&'ZU&2Np'[( ɴ8 i3,o 5P=DR0B2yCH?R҆KDi}@~RNcF2oPh0.(o 7h28Q#3_2B s'S.ˉk 7J ^GNJFc0ZE..ީ0Q8dG٧1bfn é^ds S_ mn2-&Ԟ*>q-K~c+{ڳջT4wN]s,'9\{N_M 3i}h혴:S@BT =%3wE]x?~ [["{`./ĂpJT# $cl|.gK{LvZ   z~PUesB$v ٢Kd3}c>\_D QŗQ*%z~ JqUE.F%ќ!x К$Vȵ W{*CfOYwY^(F'=p&X99)u)nA:BH _FݷW>c]koA렟m{S ]^;J ~'iiLϟ?/PrcN"L&ܦgȆ@=0I^_W=Nw;~%Lr, }D"!ŵ>xr,7G-c ]Co4Vfx$90!hOޛ;A{pPEI&u.?Vyd`.G,,b&"fz[o,*L cci%F%#=lKF8ǿ'EcsiFk_-w-q0t:(<2Dbc786qEIIvwln48d:_Hd]8,`SC^õ=;lZ#C.37Nq_}D_ uz{=VqKPo'DO,{%zF'*Zn@ Ҁq鮃%Yo!^8P&v|Ɨ1i斏Λₔ́u4IGZvGRϺuQ! yyͦ`$LD(ťB"UD Xi06'd AedJc#Ie)咔ή8+̅2j8j$[4eɵΤx0Z$9;5(]GJ)XvL͠xL]9=ӊ3Z*‹VAe oG^ |g}Θa L&f!},lZDSI9<4F-5&=_,LA=&.k͂6ȼ2&q%L+v͢,#:Jmy5zC-LՑ7'C| p3+DY(bRh @dt+lʋo'~(ؤD,36}4Y@ l* d(رT ;cL\oFHk6|{ACxkN+Nru2KP0'^9o1)ȖSk;r[Kr5qI*%غ ّH"fv 2QAtfs %G(GukeF7Ic+ʒm!:=%sgX68__A w q}ܶ#h"j-pzZ oa@ZpM9f3ܽoӅ@ͯosA0R/RNs ,U8Ԃi~*=`!#5ژWѻԣ) ro,.,gy+G;hxR!yqm{q/}/)xuR.|=*ǫ|r+Nnx?ysss˗RSʟ:z&72xy*#{ϻo8l5kjZ " !+9o]R늽/mՋ/tv!Ռ$%vnQ(6@C=2L ugǽo%|fʚyqlB6]hs*3{[}zgh$F,{KY}"j 4 /r`t$04aԍ YөPU]jz7wAvrJkUdSTZ'KbL0x̐|ĀZްÀq7yx2R%d9ƭNN1ZiKʜbfO'@.`RYٌG7+j>[-BR(?3@pZ!8x ¾^@Y$irIyjiCJ gBf:!!c/lmjrtEDrjct+DMBMQ1Ɩ(bdM\AP23vf EvJb&TRY*#)YVTDc̗:I0zMb'R6ˉQmV 1!@NP{ 5`N&NbijY"jc;MȡG G⎬jXijPߪ*迣BWc}-xS`FuKO4V1 rSFO!{#3Nur#> f:9U>jOGQMGX>OC!us"uiWk;&N;=斄;ӓq .+I ! 0Uf=hh7 i͎)YZ:9fcG~ `)XO,Utl}w83. yYz ㎒{ pE+T3uTѐ0wɪrǪGZ Mi\W:KV)>sZZ 6c ~x{)$nm[LRig/ijZ?>8?۬[z+Q> zsB ғ̢?#rKh:\+w .wxrt:cZ514Vh n8nmZPkUFBa|ڛoeozcӞZ}hבMC1A__L,As^DxkA ǬFG},>:fs79&jeBUg⺉tܾ:y۴,syn9 11ԥ}F[e$|&}6z~4.]O9zw|KNHz"#Bq}._ZK ,~aeuGfiD4Uo@L%4ʤzwn$ZNRc>Q2)p>MIOF0RœEY4GpԖirCKCZ*v5v0;`ewcGQJYA$vS`A>+J}dGZ;b>w^^C#XZ/U-Qɭ{F%0{{Յہ6seՋڸ߯" ,⌁]N_ J&jV/tQuhWeAQߗ 7^ ˮjQ|M(&U ] kus+XҖ 3B/%WBz-o8?|>;?|pb~Wru]5ϳ.6|vs_9N'1gqyrj!fǚxeVcrmNʝ @z#)jEz}u+B<~{vy3_^V n]n/J~To]r~WɾߐE<=j,b_cSDڌ9vOT.MF^$ uo83$ &bf-KmX j$laLNEsT)@ZmAڠHlaOvDV<;aF+ռA `uB[_?~6_a0z\}}C"Ii?SjPZ- 0NѥTv~8S֒ioyqmg PbחPH%+z܏`?9T;Mr$_o(}C$E넶Z)N'!ty>GR*%JUgeaY!hhuJ v}XRf/-iH ",+1V3`'?IT;zvm=[; G^.f9XmEH9?|`n@?9+,CjVhO+_%ӕ 3H)RYR׳KɻK. |\JѠd݂u-: WGW2 ,VRTٍJJ:j~n~=N.m Сc]\ 66&7 ~sg;t7oױfy}}[5a,o{FJS!mQe7a`V# opCBLsApڪ!*Tվ{#`V5TYSon_I\6rY82CNİ Ejܨ{sw0s$/;n:LGΓ6#790xjֿL#{!ڑLt@sh]Jw y}KV#/ip^FRj_i!ݶ&w8g~~a{Vnb$UkZ SV5i5[mCUYFY6|S;ޮ珿|X'[Y ݳƫT95oV2ktm\OwxuI3먚$T~OClܿC3ĞgJ`#SQs,N1)\ZO@lgw Vru˔δەF][+␔1uD͵e ,JQu0U!{rn l;Tyu?;iywp+zgˋ&}cUԊo?sd;Gsd;G~ӑݽY^\Wцw4f%8䥍ihhSZݦƲj}mhוu;[nۋB{4ho,&ix]yv1߱Vd &p4R]S Mۆ7DpД|*u}Rl{Ӿ^kp4o&2Zv::kPj,jml@衟\@qp~r8Rv̡!J!L ~L;5dN (¨ B!z"țz7I$?ܩ!kj؉{_td[ٞ@ؐoC/@ WmoZzV:[qZkmJE&(nr-ONZ[;I21yaT[lEG7d#ٺ S|EcHԊ'X`L+{˨2p 9())lHW'&dU'VQR;nB1Iؠ(qmTUqjԈ XL֣6Jb]ae” m+!]lN5,pXˇE[ĐjImLĠJVײkjC#ck]{P->Ǜ; lMR.\j# 5j1p1њ޳kʤtG!I>+jxN G(̫,/<\,tbv'wh;N k2PQOB\(0l2 6]aon l-sͮtu%"T ߡnt?99x7RJTfR=Y鳶RD=6֩qB*̢\~w^ V}':?Mgf`]`쇞PC^[W}UK,sF =YLȓQ==B)b5{GSK$w"hL:%S0y$J'OMPW$$TW) Ƴ;7q~&Zx S՗U.[W$_Fh&ӘzܫIڃ R_ưDyXrI{3C3CA!tO(7:wuy]Eùuk2((wL4;ȟY0چYP'hqNzirtwrHv۬O;D!ҰB)ik҄iT?Df4 x^뗸'IT(TG]8zF;Q Fm_ڡv$>^ާ?Og&*tM,o$7Lg)Dy\7͞zN\4 T,~d7Fdn4J \5#s=xl ]BeΣyJQ"%b }{dj^4Π=B 28$m1NU.junloY8Yb#NGo?䑥=k@=}5>Y˺Z}):Th]OD!JD_=v>y 3z4p瓇WExv.Χ/枢nh>VNl&vc)ؼ63o8w3q4έ6C4>ƶKjumj%C()5ߥ#lf(8J7bpk6*Pv1hZ4- U`{PȾnLH7BC"Wi |S5锰 KѵT{+\ね\[x!oY5 MT6gh]"ԁZbdyʤe]v &2Qv4]cT:7H\VJ"fOMFϫgeKdm26 0nPvjF=B8eYAs! gʐ,>xҼ^'֡$v3 wwfP~C ǁMp:GAذ#Ppb6n'/a΅߭~WQO*u>OUxV礭@xis\^vAփ& \%R(p3@(*%EÁF+ `3`bpCbЌ}7w}#TsVI 0嗫\t`7}8t *<[fNmp^c70qNsN[M#u`=ر4K! UFCjبRQ!hvC&+^3N]L<4*3㉫\]i9Աr5^I)mV*C˜%,ΗvZ,ǬY|<\ euC7vl`ɛRl!<'יFS!N~rpY38QN~1S}=}mO5#;+tN/2i,+XNru%{}jlo]мYKZS/8{̬Iǩ>GӀ 7 B>d?,'+ly`S&g+˜)[6 rlzD+!+?5`@6s1l#t@ 8~SQVO5ɚNmC*ӓכ73Ht96D@^X78Dgt vQ&T~!R`|2O/CSN]/&jw6_gc#n̸Sx몗U=!<3FKh2с=|=8RΚ#6%L6)XkeBuֿDj "FTFhdۘlT) Z[ a MW^QBaQG/Gu)P7.䢃}D{}vg͖L.:ӫ Cx@/?(BJT] RH{Ҏ|Ͽ&]^SXq*i diR/ Lg&CمC7$:9[?^Γzݧ}YظgU J|Үr\)ĤIjp7oˋo;;D O7LE]4t Bђsg`yy+wʮN`,Rb4M%Z=a_ޠ<2gy =D,w!,~YU^z\=?k40KYM8n#b̗bni,I60(iIOw=QGv2E(n,DVUuyYNWClă\KHu`lk"q?F^!Yc/fMwA1:4&1&5Մ1Z: ƪs=F'SĪs=ATB,k zIk4;zpFXQ19,~ԌabeEtLۏ@Bʺn 2Ο=SM 2k^DAhJsKOs]϶cg@:u{7t)+;n H]/X$*.U6Ә24Rx[fn7ic,ޫ;f%RPH~0yZdbxEe،b+;=GgL\mlWk9nQdP0JU2d mg%!`%R=5 "Y\ ڿg1=JHBwjJfSYROa}Xr۳$|v?:c 4cwS>p=Μ$It=&|[jlzAѳ4ʪo-LBKI=ߕFQK/OKkկnFvKGxӗzriԮsۗb9njz;f۶s;ivӴE *{Io?V?Vu*ffvy"4nR T%+*-[BeTòj2Mw 1N^~C"d4Ji~O1%xc `vELAL}ϏǑVn9tF?Wzp/m\ʩ%qک* N7KS0Ud^dWG_U\j6%ՆnЇ<.|Wjm}өl-R<8}`pѸɶnZs?7P,+x㥣9iKn14)LsEX4/^@OJrR^KqQ$ g[}sO)BNIphU!.E\٩nJ!jŃBr>W J?^5IU<өw"B`FfU7ӣ N>L9p&odMulrKYr:<Y={' 3Rb&AZ~wnJ;h_B,ս_1 D'hmkm{n'VZq.Ia 4cO$N~ڰOA@ޙ>.N K݅Joj%HH*qoWNܷGCH>jycc.2>>Q|19NN> d1Qbr)T + -f9wIP*`ڨt0u5QK%S5&Nb0NFEJw`-#NH+EeTFSSsClaQ,bm+6M(BG48 slbF4W\WV̿rYDŽ M{gsc>mqJFЬsfeIU~+&X;eIj{ӾkbMFjD>׍Ū:Y*ޜ?NVǛJVeomO4N3{}=΁(<ZGan6[R;T>v[Un)6BsIn*n _nSM QQE%MQRHuTfȮzSnnSomkdWPsG3wJeP^6i3:̦ʸe0Z[!8{Z*T.l=Eޞ+//-G³9mvc6}f7`l'Խyd B`ogt*A%S_‘.yLDm4E#҆ 'GŚ G##&Ѣtzwd,*cvhxDK+:k48ID><ؒ8EF?z˝ń⚿]m gOr˙9vNH}i)o@k-0- Z1VS4f0EB^ VRJ^e^'(ijVTr,K xn$ U/rb0+lP&64N힊t+ʹ~>T7?X'\UI?Uiz?Pݷk|~-&O_z괪ovYyj+ ͹篤(ϟ߾~׋{3Ts*oO>+fUt`RWkۦOJԁϟݏϸB^Mn7psU lS(wso?#n@1Hvv W$ѯ]J/_)LZgUKaҳ$]|ҋRxi)ו\zLT&^f{d2V$5Ng$C,Kɩ3)p_1Gϗu(ZImQQI 5Siy (Jn!;3ƭr 93R'5B[ -`XAu>1373&cL3 8SjZfR fhK Ҕ,'Lh(LҞ/uuƁfRwqٗ*?-j&z#&6ek)΄䷽zaa5P߭?x=<)7qZd mypsRpoH\_vs@}cR(P8^;"FIRZݸ 7n ;EU~'M% ɴ>7&E\uĮ T9hvwfP?wS#s&3W~1:ӎQ7d>|wtH}[R#7@?Ў]!(){F H+Plت6+[c:]QQmˎ oGEރzyc%Xn4jaх N4.M Ѯ>3}Fr;rSӏ Tc;J$f(4FW+00YuePZ SND*{0rs7HmBK< ?=X&Ux5zsU=Nfk<:͜^PCE&ݛ;vZ܈Br4b<[TXʱȩ@ VnM$*_x~/T5&ߪ2 }IUa5ᄻ[y(9OE㡵ȃ~kBT)09Gh7{mޗ%hڗa8N>^hRv*D-JI^VQ FYQJ("cHN7x/J ùGD>ndRŮVV}jo WDz7UlM[}녯^xKFB;8Ңd5hgQ_Šӹ D8dB0((eh<)PׂJ^ʟ-Vp8 6ɏexL?Cz]jVy{]}.>?Wzp/e0 S|)Wy65SU(b(d^dWG_U\j]R4oz+5ejly[Ti)3F@rщ ftApp+Kw\.v+U31;Rkw3 xF̔TS<rߛܫ=RQwQwrFx&R,R|=Ǔ={}n (IIm˓pGqssKvAP)_=.O4KEixpc%ӴClG."1QRwk,4&8y G-c" l*J+ӧ c@%Vj G28h]cq}Ⱦ,^rg_wSeo(l2iݵ΂j 9w|NܖG6-mUQ<Zͺ>hGu.bRQ)NS{xw$d/U̧&rd;ɚ} -vf7[-؝{Odf8C:4nncһwZuP;i"qj,Ky]fv09A[ joD RtT&D }eB=y+DP#G ,kvN|z{"foG ;Лx|$ogq7ޑJ1Tmei5[׭f%PLgD5V9ZeOjWKFzN]?%@;)c*@ MXwlFU6rHЁ9A},q,WЪCN?'X ~n.3#(l1psٛT-r၉RXEqOٸ\VH]ǿ:csyBwl  ќi[P19/ EU*F)84*}-qZ HiQ8j"IETk񖥫Hk%dBqk3֚LeF>5";< @ Y YV5=vH~#d1Uko_ǂ&Bw}2^dmH-Q.]Iك(zȽ:h9$M?TCZV>m$ Ts\Br1^m$d8h4LYF6X$\Z"bn0*0FZҗRK'uqL+w7ǫ|qG[7/S7IMҩ=4A(WY:Wz;!]r)1aMbu2r)2>="STcxrOw&z{_B؎XJS仹ٔV_r,Jd" H+g2?{n0g tՀXgVvqOm02x쵫j$S`sH*!4,){ ^Y"?5Om\ U߉Q!hr h5i*R>ep( ?ai4@69fc[ZQ˪BilAiYE޳q$ _4(mBо>fy/zCY!gbSaH0T/k!XZp"!gB[cFsbBTdXx{}Y7kedSA$[_S 0UyܧSaR4Zsh߅5\Zw= lߋŵ-6[rBےZ.$%=CVjf';?$)^x9*&KINjs*{~ K*)/?l~9X~DCp?J9dB=ڽa`ID^>{#|jyzqRl y&ZdS}cw#lmѻbc:ݎw%ͻez!,䝛hM8mN9w+ tJŻs ͻ_nCX;7 ,U0K{ ?gsvwq;scz9 Xu_>^^fMէշ/? <7) Tk=sWvgY ds-'ַ> i5%`bSmDOim/ew;TB*٬!,RI{%7Ja[%00_q=(鋏?$3 牔x`ʂEI˫5]ȖY#wMD~V]^Ev^.,K3цRVxmgOEdljkڅI'1yPEɞo<Ǿf66s;wr6gg:>LӛnzFY !Zu`TI vݙWUdVs|&^6U:SP2vuŦnX1uH*'\/JLT!BQE֪*W^Jhk*24l֌U7v@NSZ/+sӎ0D@+ƹ̡Zv1a7/v!7 Ǽ(B˧W\Aùb rQFA}MoBN.W :4[!BG)QBZz|8 Ώ,5{)#q6kUMM}),X9YeE ֗ $RK[zu(~ aP鵜?[A:=]voE! p#Nz#Z2`i0 /6?I%YkO(K9;Zܼr${@QwY*󍷠. 2)t8uI:G:=ϏASc=zQ_VјwާBjydީw<4jD'`rKL{nV"OjeZm͉`Ԙ4?M͉+zMlM>X]lJe4GARi!2lNq"~|:l5inC85g сԾ#HbLka!փx)o5?{鬱Ϳ`Díbҹ#eHS -'sd/ i^FQQ r *u`LyQ֥Z;N7Ԇ6#,j85*66Z $#`Q B4>E e#>z#BPȕH"`w$)4) ͒d%K!YYQTHTiADj@%6%i8٩LlL npKJVźD2PCX»iU5\b F]rvh4c0=Bx,Cƕtp0_K:šô޳<hjҬW4iAz$]xc~ޤff LwWL'Iw-{;|ztS-@:hTbJ6j=kb|b:i<?uR'%iǣN̠꛰b2|ݹ#dA;]0$!aG14=U_"5-ښJgR=j4J7~VQ\_j<}_ݦZ?)9KѥCSk<?s-)l-ZN5Roy7^y ޭ)Fv+TxK2MȦ2)qn{7cO»bc:ݎ0n2MMZ򶞟넕=tfU)kkT?-*󨹼wug~$OIZnϓ:EMd꒴wH{dɇysl Ղ\6w/-ͨ/w@کHLW6õ'F5Xo\~' },/@~wⷍ3x}ǝ='&"x/Dي&4J 5͒jjoW̗^.ir մu=FDM ZD:em oU03URKʓZ.,H4*ns) Bゐv#iG٧%m<Ӻ__]֥\!-U>ٗ!H@c3Vh,AbZ+uCluVț5iuusӜ^7pyDgoy8n!6th7YZWq] vKxzZMO/;D-)zn]`V d 8 Tll l#iif1ʉBPgq]Oʭm%|* f\:PSIdҟ]5_2Ԃ,Ҩh+VCAu}9 6k#a[%Vɨd7qg* visWPJpZg T;xFe|(ahq|2BAQp,e"Ħ=r eZ>^PKU:PLL(zi=4}pU d7^d^ٺ&>,.(i0Ծ~:2yi|M/ݤ &Kl S`dʌFD QMf!+"S#2zI6A`b%JYo|rW:(Ů^4RsMlj,4ַZHVX9 8P5 (JK}stdXue1 K:Mi}GM;R izl=8"_? IHa7/v!WR,An|zK"5zId=.e) R:i41÷Kut. Kڂ} g̥9vע!' }z3$hz)/k/Tr#Q4.5c"'4,sVK.,Kհ9%8LTh-c!e)Ì- ݃Qx)trjrETyKQn(hb/*67lR}[_ֈ" ]>1$Ub &x8" rdw(Z NN)_m#{=<تA:1k؞`eTnUYJHoЂ@r3M3q;Gaّu#`-9MaΝ[.naNG;Gj{5~v:+t,teՊJAS~=M?~:IB.1_l r۟vzew`4Y_{nwyRĠ|5( 4j!j\ a.,Ț ,,`jRh|g@+. /ׇqݿZ 8uj]DƠr]X ޥ ۥ\0ꝏi  hZW{Hg`%+y)q=='1 S1M_,^Y3;vWWQ&k /9¯̳ɝQ`MHLqz &\u^p@ي12Y/1FP(nym~uhF".$чNjA ~ 9cjb(^.—0bĐ'cL,#X)ӳ:gN'zn5׶waG_~$0wBI0 #$ӹx;hP |Nhu[֗hҚv n1F+!cU* nԥF-ۭ]ݫlಆ[ʇz0.DV7ouV7ouS[/8+oXcdPHM3kM\$s臱QNf|vWҬǚqu3{{uQznQм^yϤBȝB|:#ݯNDbprшD`2: D)+'SLrdoÊ\fLHkR#䜵XO%$-T ( HCݜW5Oqk%Ԩn<*yY>P/FR( lŦ2,zIscʂC.D\& :Ā;J<[  `o:_ByȕFjBaj dr ӂ?_jL?3+ RSg~w^C\D#6!$vC~H-ynԜ>gbsXFtTTƚ>7<ϳcW>DA_w@̘[(>C':ڭ;qbNxQOc!2p%#vv ka l.BЊq-a!z#Ǒ|ް7^%sC/jߝoG_$jnZ?2ͧ I4w9ރ=#Y%fh.i~4vo6"" `B^I@vܛ4k~D-}qQv)AZWu巗PRPI Z*۽D " {)-wa%uNIR,ER"{6BHSU `B{3SPRk\WfwTYa噔=ZMXtmE2yךBTsNL1F- ROjV{ Ksu}!].#肝ӑ_[O󷜄(%Cnmg/ 8%'#B&f6͝|X!zDz IT^5CHT(2 a Y @EWD\&[/~L-8\%5w,mhgx6^>,,W3?™x5Dq?sPeYY 6v*kl.+"MhX!8 %VV8lve EApdz.S'7~-"ss1Ճݹ=;CLq/" Q,E"Bi$%~LD(iԌٗۡȊ>|E]0rf!3XF v:q *B"XTrJlU "u*ֹĚB.8l'h(&>V=<•ImpK^TbC޶G|lz~=Bx!xj]ϢZz`ɏxƫFVҶTm6Ȝa|zlw٤i/FIq+˴թ*SJhDCi(l ʉy{fwuy$@5G9Ր?ųi~ ]Pxw%LZ9 ,Yj@ h)3 R!92 N E&R! zwAp%Cf$bǡ[է jFHDβ2#Hu5 NYC0U~< { +#C4RhյZ-^̭K0ǙΒo_f7bki9fQj#c1~NS:(; 4|9SdTFb ѯ+߬$Z |"4Ǐ;x酪Dϋ!#Q]|]ms#7r+*}n(4*}Rq.. mIJ&Ɛ\p^8JK3iFqe:u{'O[3Q{WH䋆aϨqIf2}L(ɂ ٞw8Cd?u;30$ Gq%3 7Gq+;r5H;;#|!0Z 﫟|{+ uo8?MߕjqbR%n%^N:)->@*x &ܹ V( ?:`u"7N7u:W~?;枂ƦM,)`+i @SyLJ'VA!q|jE5ֻڼvSP; O&mk̈$QD.Fk3(wBerqEf9;Ŭuk8|l=GRiCLEP>NKv5)Ex8 d78}Ao VWS4YX5ZKgK7WߏqF䪔gIef||ZΞh/onndyG㲘]m(S4^@;;=-G\=ǐGȠ+7xW'?{E^~v !jT@;ShXVgZ%'XxO*62s5\D.DRWZ )J>wS"^^}33LgcXo LHFihכcp>KJr;y $ʧhHVyc3wfZ-f]\%ΗXgO4Mf,dB>R6?!c9/_Z~;udبO0ϐ~ac^6C!xS{1d)xY-'UtM~à6 gvaƞS|ݭˊ<7q!c;1WDGm?Z =6I:'}Ij/g sbTŲ-uz DiR֢w^T3tCdMwI;\s % T;iJgc͉ͮ8v!zgW 6 yvaHjXabڞ-!ZɔZKC !>s'6>-\9].y!8юS`-)OI:E#dNu):LwC}[gX.i_XC]#9%HK,|ۃ5Z$Ѕ܅Oq(ĆR2ܒ^9NeY>%|dZZ\5"KWSj0L&_Zi#cZ@YhI yWF!=܉lRꑯ_ ] >){ D|M Kk*Bes:Z 54^6;]*)= Mwٶ^1]qh+s4=KdUUte^v#)l\=Rr2w m{L^iTrSE(nߵX+KE-yqÝVO~L,K1.4 qA{rT YQ AAQOs-32B룥i{|Wq畯4*+Y|W[dZKsB}En[*K5YJSRX)(Z&'A;} 5P@ 2" Fp6.ՕC">,y^L~ ?^\rOE%J?˃oc_Fѻ^eoWBJ};W?x| Yum+]^_|\d2~UmT!tz+嚌L$_,u]_4FĎ)}gWkzNYcy3hl?F!"?b]wCi֬V?S)%Rnƚ llfc-.67UW4ƊVv/N^O^XktՋOԠK83nֱ9 `r~VuEeUE Pe-WrNu8I4wgr *Xu9&h@rHc)VUuo+7wOSx5٘vx2[8̉&~٫qrcGfil k!AcXuDz﷒l0&45#4LrͯUrFG':o |V{_KsUpG/oK68 g?76Zj!\xѻ(eVS(_:ɷ++9JO &m;}]eJi(Dk9M(AXzu9os_lcy`(x`(P?i 6J*B.9N@(av0+U0; 5Kmd &_^WHו;_JZJA N:Xԡb2ӺR D%Į(i)DmuN$vN ne/7w4OUHXv? IY-$:qVv)jq>/bʍ+Tr& 񀁇ePył qu y7Y&bpK%v^ xRhwR3`rvGkӴt<'FBKI}-5H-=k-eVӘjj)iZ6p S8o-2MKs)RKI}-5JEKQKE]*Dq".ZzZj=݈RJ*Yji)ͩ͵^S.5Z܆BKVJj06ߖZ'J"`Aے::E!MKi:O Ah!ﶥ^VRҴQ|ӃRҴ̵')-7$eYvԅ|v6O`HDDE\탘I}-5zxS_jAd`Gܭ$«/SYՐrMZ ΰf!9 'J ԈMđ#V$4. aTk/%vQ-YE%Ԣș9Zx-sy.>4]kd+gL,0[i _M{UӇ|v#sѐPXdTtSJl/NߝpU)ObC$hx \2? F0NHݓx(R^_qn`q3y,I?²c /pg?7-6?}JlOV6ʖb֟34j_x6*fB0=P]PZQtW|p,Ls?:qtM)d!Yjr:'D~B)kz" EM9j%3rs=@ 501!<fp9pKo1F:Bh$N i'+͝ RXڔ4U'H8bNј"!n9L֙֌G|$ Pk$:Ft\ܜ*lwɥA} WKKq`V߮K@l(^? OE!9܂p JHLϹ/ h Zk~ϖǵukK[w9$ى ?]:GEG|uw^+]u'٫hxm>d x?=r y'77YsJ{ooå b,7n&wH:Í<]h=MF7w8l&wȽ򳝀nz` A"g :OSh r̦y.O'X` A tOΝMfŹAVogZQ~|l"ЕE[ .{~ۯntٿF_h;>qw~%?Uk!OC՟dCܾ4&>Hn_$|?%^n\H+>ëwC/Sg?e`0<ll %5\Oڌ<q(F zث<3ko'|q}bt9->e,B >z~o~~)߀niKdK{;Aa,4qp:1W=  3ֈ!KXK 7۪/q&_aO[OQo v{CX[~g/v(P&,3<ɢXy`{&%O@y>W@r֣y)U`kP3fwU"1 :F]z>CyS5֨"E,Z+DѼ͔)V6."9ئ2&qKDS_6;eɽn ]zƩB1S+6)"pFfsba8!Qi,K#8PxJfmǃy_"B~~l=3H?{[U8J8|,yb?HDZ6*K"Ge굊8vNIӢwn|3~7#o7x/7y1\NO1E]%aV4~lb:^qA$X@/SK %oN)$:}.?޶m97OT۴j ,c҅>Vͺ`x U0gTHGZqK14u%O+\6bJ$ >~eq  E?/Xq(Dkx ‚iZ?Ҧ"m˭ur[j\fT!8dy YBN1$8TEp)JHp1ХaZ'~5VC$`56F"_CǏc5&B J(bLeIee T95)ZC]6AYL9V@Ƅ0KU:\j_)x)߯)9$Siԕ\Hay21';.s-% Ô4+S)$S0͘#wU9b7.YX{g8?kĪT xK5+_,97W׶F9? P'ԭkdY(sl{IYs.Y'I!.Lq(*O}'Y; ;GimOm6d FV pZ؈+D*Xk; ˩~S񭡼kD^\Td6I0/}>IM!y,Z{IrwwV!Sw*N@SHk[}ݭ y,Z$bhOwv7ڇݭ.BXUv]jݭ kP*\B|ar|!=JR헨Tcix". VcĚSۀ;n؂X4 l?pAHhbsVPmjᤑ"6RpaK6X֤z,"6ˢlڳeh5+j Kg#>5)% /e]I3)9!J9Eԙ H7d?.R 2 >j,펷.Ii+II@1H;-Zp)}i׊@KJV^ FKSzw]"Z}qY>L5aҔ6K!mR@#R#6T=n;ٌ6ZJ!'_! po:EKOe$FXk2JbV-8ʼnƒ PcSHr-Ҙw/4,2AXZǘÿGVp`#e}M¬űHlԺp( &+-F2@$ݱsvSQL0[ ށ E(Dt:n@Ǽuq5ACs0 tDQ4VKXѷ#iR %mv=:#8 DcҮ*\\}ZTٶKX,=]kkcha% nP3Ib@iאVҮ]W+Ԇ]$N eRb8O8,%N˒cjc<<4"l jN HҔ#\GїѥUugKjJɫ),X 0Fw;Rju;Gbыܿ8Լhb.+ǁs L)M#g)"-LUc̮bI 0+yEDȴ6 AH2%DH"E*ޟ*{Ouc|()U}C*$}-ʧCqSr^+GP"D(䁷f ۱BX*G"$rb%:]JTjSaåMxeXj)LP і=]W몍o5PH0G;NpB 521c*Ҕ%NFc !Aٻm$4d7YMܧ] $Jx㱝vWTw/h?MRX~T)y;iPsЮ2x\y;-xX)]'UA\KWqt %SY䅟hQ,@aE],h9hhT*]кo+8b&y*ץQW.sE QVGcz rߝ''OBXcM9qvP>[1J0̱;wx>50#LZS 9>1̀:Վ5{4H*%lq}l@m%~^fZCfSk얷^vGtoSywjvs[#W\[b՜8f߮ܤ2lƳ]SV?Hu}Oһfge7uVܔ_ޥ0Z={7ێBk~ջ7[L0c7P-V%$7c- `AٳgϨp 9v̉y@=X{h9qh{/S0^ ?7aC'`/9 = tZf,Pe_phm cRǟi\}~ pi8f9v4䅫N@2sPj󤱰S&16l—r{ZXdkk`(u9yUxbJ*y(݊䯾J#J>^@ KZ5ЭA ?V@\ǂ7/Ln7_8/CN_ݬ]kTq3 Pv пSvu6ij(^ugb%'+N1f{њ~1" Mé8~R{`br@uQ"9-ŦJ?2MGwf P[U"˥v9a.9y!+!JQ<ލes嘪>}=RLTsɵg|F,E x|K]own7;Ꮫ7 LfA ̡Ҁ?stG1*O!>-{DT'.eqٝBR`$!ǂ?,AjAbX=l {.J JJ:SLuBׅV9YpQr% rfPnz V}kmuΕAkML鄩 }r` .d0e)\^ql)d`!Ŝkc"evȤay ]u^dx7߂ԤK{!_r`"()ѥXAnu՟mnY'%kU(B02_Ne^XqW*ŏU}k D r2yA=a_M³ `юt*p 9uYş!vF$LOZm?HLY ݊YmB4ցV)qc67@fAS#8`y6 [* P\8(̋!O atVJH U8%ӦE.L#ȕ5iʅV7XV @Q2T+<)98=1l4sMfۓ䅟6g1,0 3=RNӵh`QgG@r\c 6?+n D#k}͗0X\1NY3fj)DU;rB95-y[e]^FV?p8q185(q"Oqsd޴T56km+Ya3^I9Uwb~υ:qnx +E+n}=O3[WQ^8*C_lt qkD;}:'>4/fs8o.QиegsS`1:;[@YsBvC(m&:|@``Teb]E@zW@Ysw;yMNm䛒Z7s, Qܤ䳧$@0'cZsM=|%Z`G_ 뿊oJa5k'jO3JZvWGq2z!Q hÙ=qRύ.}ߑT鹣鹑ј̜Q"A`l2Fc v6t E>챑kSwǑ,g37+aϠRZf쳨yř07)w zx(iin7呥+L`\?=*'pjfSMHCOVD 'AD@m:@ ,e꧚iøX<~%jbzԸ|0p[8aq"Mٝjg$BP !?%}5sJAA7bhXg5VU4Eg&X7r[*BT'1mЗdjx-}buCC^&g1X7[SZDN7tY XӺܺ%KZ4䅫hw2Z 9] v׻ bҲbQj4rynaHUh`,d9͙ ^oM斋 d I9ͮ>\<:.rS}|(Z._?٩{KWOX^*on{?~> ];r~$?/?7-XZ=SvE$:vK8lz!Ŕ.%նI"i/K/HZ 1S~l!HO`'ɍW;ƫ%"o_RH^(͢tiq5S*Ʃ /Bs:\TU6GR̾Ⱥ8'D77姿ÛWzdyi"%sX%DL[FN/.)n6A,nnV\6#S-./qqsE_1QL^aK+g1ژLa2kfNYM]7VBgFCta`aQ=%U:HR6{MOK2။s&RjrNftd^ XU1YAT+EDJ U θZ9Qro#GQ̵PFgǼdA/01VںrIl-LˍsQ^rWȺBVNsJf>@>9^j -xODJ&2BiӪWOXvНS/A1I|әR8^gw&}t)MBámk2zi(rQ7KLuHss8GQLVͲv}ҿ-C^x!gYSKWd?h6 +`É%QcYA2ЅTd#eUg@ZpITNDc +!:7h^,\J\Y;0 hY*j2ߣZ0[dغ?5ղ$~L 姬iio?5&?x*'z?_ GrlAނ Sb"/ju{-fo~a ڷkf_Z^W>q{O(YqS~yd|s2HQ=$ Ź-[Ȕ)lf [ׇGٕ?{O7_1egwZ> _ n|4 N*/m\uEQ&H^7"2 : |#.">RV[~R.:Oxlaŗ^%kDƞ|qm&N rAxD2TTr-P1gЀ2! S^ +~v3Ufq]o>]]_5XHb3Kvһ/tLdhN~ݯRF7}Ζ8gWwwy=yqw$eo u{>ۂ?ݛqpbh~ Y>]|e.A"!Fl[p2P:Ħ$мZ-qEɃ:/8(D۞rJ.jܨL+gRk͝4:ܞz.' vQ$66j5u%vk֌ QlqRt!uE~t5uvңRqR1[*۳5h')=^)e$.jm/I"?gqgxڋMh) g^_{ 5 i%ڰs2hԓƍ?$d`MGId` `@&l4j<.PU/:KJUBG IWl?Ķ:fCݍKP:6Ϻk]$Qäu6hkji\R9 .09ϝI,AD[AZxYxcbVbc޶5n 9)脜~_xN2&-ZX2a Gl2(ʘ׵Ƞ쾉 lr{/ў\{re'U{RNnFq<Nrfp?2Jdj"8hSs_ &^B#/z}V6BAAGl='^\_^)tA+E={I9 .nTL(89d,UI:De$8EvD7&Kauz{Aw&.U|?}Q®vUMi-ݤc̶=dcNḼ^)z)뼨R)rˆPURY$ق Rs-WSgr~@D ^Oz?qjFjJwƮ=r^fP1PҤwWqTeiG]\ݩ4h!1Z#C~/jPkNRzRLdR^RJIK)ad-I)# .W,K H0JB+WZ5&A0y )P䐃ɸyZuBZQcwjSP짴ʋUvs3YIvE7]F3I3 5I$8 P>`~nܷˡ{!?n%I_H ՘*qkMYApxҾgMhG۷Cu^FY6#,)1LKh>}z۠yxǜI)5f՞M̳&PHp v'/-h4i* g9LղO>9g*F۷ $W̘\3=ta@OAq ePb]AADC .un9ԯ3:'/Wu3 J/L?:Lx=zt>n0E=YƖG&R21=tp2>ӶQUdF*H}mw_ܽq|#&Q᜘ɳ]Gg.@Fg0b'c/C2CuKoMhxTpɺ7Q|I*ZcBHڙ;`w&ذgCb07:Fo>vPp̘L1(ф]F$CVF} kX߂lUQ'G%Lyd1CG>pC6^I'`=su aBǫ :7%jfd<Vכo^Ũ]Ldhueh.rSnmRi@͍PmUwj _!*_H]'U](Je*V@nP@`F46eb9#,4pE9.T ªjԁۛ*)PJR`WoUATlVZ*)5TPk3JkoT1ʀjoEB QVZWԞl|B EwVz&UJ5)2\J8ܔ׿A=1[&M e)m}L{ssOZ |XG"4rhpQRjn ftܛzXHoJ+KPX(fL~/Iv`R*1k# bH #}xԎG0Pxӊ7ٶڇxa]C}jA9n"^ū;vk}5Z{G7b,g̙!Nk(Hc b"4΀gNeqgz=[hEEfK(j(r)gD &2Gͭf,NfVpY%ѻ 7&{ZvLWr#  +4w}ް%'o;\i7u{"F[.~ęƲGآsk0z):׈3C$#Ɇq es}҈A1;c9b)t%VΤx /J7B_;g w>;Z.t/!;-]ͻǖy,98'Wݧnc{­v2(ePˠ$U%Yٯ Ra>{rz [xl퇑|nb Q#њZ}_ DPywy=ٱa`8PBqD'F0Sy [{d$r{yX'Ϟw7W,9|4*rCBgC}kfaCjKZwk9 !'@@C0G1$ /npQԮmп ON$z*6U|!5=ko9ȗ~Wಟn8pX0_ p<؉L2Yߢm=lՒ3 nbUqDdQʼ́)/B0zvkqjyw]-Ԓl6xB2 8A2F&9 )U~8B# #/<x@ ;=s^5S'$*sl%v""g -k|!"bbT b%нCrZrZrZrvTdãȹTH1gu@TpNhxD :)o?^sJQ^ŅNy ֥HmW[/]\]/ߐW)n‘}>X4?S5)=}Iy1&*3.8􍔠 H&f́]YDXEhTNBϴ:O͆" ]BLvOSI>OPz͓|y-I)NW;/KA% %ҡ,xMi (cD1]µJj:!~囸:CD:p< CJ^(F*L4lQQ;R1v=Yw^ 'vJ.v+ jWf}Ǣ;im(?A,e!nf 79N\]tBnVsuJwuw.jP+)JnBfh7|f7WsntTv۪ܻMVf8,vf!ƙ I1:Bst0QFK`{=A+#2dų +iuX/2ﬔXޤ5iӁ4ޢ#]]9H>B:vi+8ÆDucROk\ݒ,O[P[xCJIR͵?F7KX'" 4.:*&,D5d,|B=6 Ǎimuw7A(%-h 4X"ӰsH$J:f"8G6Ed>"6$wamLQkFZ+Ӻyw-46cZUfUG>eY1LV RPBz JQUl"sTIi5A '#xI:.̀ܛM-04Ri5Txӑ8#r{W.f[㑬GRf/WXzN:(*y53\Zsɳֺ}jy5K*C!6{U Y0b7@'^dC.G[|9:X3{^N7δ˙}ҝ`UizWq\p _Cy*z=2-p5!G Z'vEMg2VW,`X2fplcrGK2ύoCzU?Lе&ZE>eP 1F΅(6I(u~Y4 h}@;i h_bnJQe5 Ă!WE4J6)lpI +c@l96(Xn9eN{#/v \ߋJT6U wr?^ND/!ڨ#\sS3JeE}'S4"(ۺ?ƕt]wJdnew_N VMnzu:AdX :Tθ[_pkxIZL,quwk~j@&۱wH{!Ywעh J,>J | k_\{s'|oV>/޽|>}(XdJSR n*a7!Hw;1$ץwnAr'f׃G9?bIg<\hw wSΏUzSCQ|nC7b`k;ZR 1r1~zta #Fq|6ǔ>R<; 8Q.PG9x4ԭ Q.\{YPLW.>_otLՠ;@buGyk+眀C{bIQAl ?w1)6T*2 K53d}p.ަ+ʴ<:p'?>uH_ǞwLiGHޯx;6?j"ZK%O懧1HZcU*o]OL&O 65ǨBb)!"!ev#B̕X85 ͷ zN0󹞚`X wqƹ]5 nOٺO8µI3HR退\=fe0nUuTm1ݒL4DLEi0&N4H6c4)ntM*jS{i&H~qjqIoJnʼ2a& #ݞ-wjƹk SRls+?Jz}i9 z=dT۷Sxm;At[It;AézvXȒK=^ :紻/ۋƝIјs$(J3^T5ͪrꌡ4Gə/.pQ~ztcqH)ƘKyح [Զ: NTwRmf` =_p?FFaKP%ki7z\oD88f1+Ż]C",ӕ*i7o7? ~Wzo6-V"k/*v? R !!r=XPV .jT5ꂼ#_85+GpU[ -)-r`#'^7C%mDt:pLY&"!]u">Nuq|n{S܋v8SH8Ҷ}Z81!Ug=$[LrFM{Q9,=T8v**a 逅Z>Z :r7z!s)t@Th!)TMBu]hcWGgn0QE~^بC:ώAP'5~4-J mݚxwֈCz{X}BsbuߠD 7cѲT|J /Jxi?>۫۴0]i!PZk "eFL{HIF 2hNlp'1qjCVKP % &]/`e?}ul tL%_y В닽Q&G~[n#0!:-6"L-" !$!'/7WY)e(/}%oWsbIe)~@?G ̻@k>"їMW 2}Xʄt=@&)HkSL`r`.[Ϙ'l;Qؑ~{#dusδS;= ϤNS(M/2V8ydo"!R >:atz3:r RSF,gS_|ʨg+Ϣx՟g+:&b0{?r؀2dɴ#(MMFv_8Vcf/>1N66hd}j>ͦeD&hG0Q2/'fO M xIosJDp'!1֘Yr3/S\Nl2bǻ(YEte{ȁܸZ/JA3-YJMBVI<gAr%@HD[ 3E]ysEq Y Hc6y`)4 QfAf}cMk3(e;WR4Bv~62]>Z7$fCDl0&ٯy6 1'%%\KѠshG=z@->:,< I"='ؓd@0>-O{u|Yu/cI,@} A `ŠlIJb뷚)ʦgAɜqu鮮ӧNlڴro"jl&!*RzmY[ 3LJ5ug]BGOE Q#jɬ&>}1WW ^H-i 9B/ƫ8g lsHU vPlb|ɒvl%O5P*v-i oX[h+_CqNSRuK۶k WmU&R `%Gah=goU ͞z9{LL: ڲ1( H'efN?1.]ET"U!M,+K`\VSRYwOAV!MfJ%F_oP<P8r6 1夡z_qfԬ V J1z,ipTUBP"Vn` D޾>x | ~D7>tE4*H̀YPFxIAWK?ɾ$±3ݑo(RE,pEu)AcZC0R4vA]~^?.R8wV5im4^zqQ3v`  udzdo Qi:fpQn9F@9ml}VNb X A|`g:: X1KK3O# g=]78zy)k_z0d>v%ܔVGMǶԊIۦ @ɪS^&qd$-,WbUO0b]')>6DdO?-fm[t>\7rT#-XO:d ެJJ51ɷ8[Nm# R7vyhAdEzN.Fqy3ivtVoLtU9(җ@%} RJlj?? {- e /^0L??[4?׳piٺD걛/nH ã]u0cO`Ym; 7UݢMGzn~\W탭9ِr葞7&lKev~ᅰջ\^_ ?kS#];=<su aw/J)߻Xw73_ޫ6M Gا ˡc'XEZ 7 ( I/ȷCYz)r9dn=)i håLɿk?=7b rs,?m+\=Ïvߧ8ATLE4*KY=8Og:CPy&piSw~=smsf-0PR `:{CHqJrgzxʝԈ=ᶏNaM^Q]JV3AH" áu#PJV)DB!b0I ,S7;;g4^rwI976:z4Egd 'NJ^J K3}ߧ*$,/,\hf CiU;OM~x;@hhN]iNDs4>ќr_aBszԌ$4b9A۞±rqmwqF\">{I-j4m2]=:w׳,C`tCNC %~e$[eepj!_۵*8bWxuv(5:(`F)"YSQ,#[a>oK ^܅i6`e[!+,`jEMH0du_GZ%娱BDBh{v dͅŔblUZllM4}@KS0ld!F0,u{сYR:LM{y6rߥa~w3\ ߣ{a}bCߵzi5Xƾ>Vz.igC.PUhmkv=VHR΃b@ZW`Ui˩TN)DC`KdzDt}ζv$ QM: _,u~%UeSX+:pU0vt#z⽞;b+sk8UD{#CIy,HL|#Gʾ2|3w F.-$n} ]c$ %A锖 '⧂ {=Ч5K֭ZI+Vֆd J׋u:޲;WWCxl}(H% sIx4JsB-5?=yo9 i>;֪T<ȉ-/=vЖcܭWz~)M:7(Gq7vl6<Ó*9A,֟؅Ml i6uEqi:vz}n.$;iX\OuE;HBX?uUgjA"ȓ߮ v@I%Kc邝,?oJrd3]?̺" ETk$V E3S%_Ƃo 5>C[{BxƆ(nz'*ȱJꌅwd!i߰S͵fqsBk]l0/I3Y%ZW >Ɏ*woâQ 7%}\8V򡳞ͿQ"~-Y-azfhP9N/}[;l^9%5zKհġdo(vPEz.Wn_!D-^;,74Sfse!:ɚ:YeţVt24yhGy B !/Hi-5d?4-t'{jk\//cGhUeE3P&+Hi,mQɢ疏7wn{#+ò.:,`/tVy6 X8"Uh Gu*qojUvi뾟i甔pOyZ'Yo6F׳PD Z6,{P9ZZkt6jXAY"# Z@V(}ɽ'n4:TIbF iU.'tQ!gca7u+ϿM_>r Ke $*mKiK'Kq*].QSNsZ9AH;+"Y%=P97^&`/p5)5l)~2SRgߨV#lM!]]?nwOx,;c(/UqiChqklY" ;W, v*aDʛ<_, Y@$XVY3a'C~pQ~o<Ɛky鐃!Λ ~rr= Z64Eez;w0-9{$9B/wT`0 fvfAuc#IU()U:2Ȍd"fx׳Ɠi\^ϯGߎ 1zH/;1ښM;:ı jpEE ?qӔgbG.3i&oٻayA^kQQ>4ӣK"|e6Ռ1m0F7Hp C^'K]OP+C4k<ۓv}ù"A b=>R.= 9,ZJ2ta6onV}6<AFB'@fÆӵJ!,+iHmuI1B@uq"㦢i35L5/X>Kxw mE)W:{{(Fq<YbiHCJrJ@I흒wx#Ve n|iHB&3|5?h 2]ĔŘ5FfxXWZnI3D:jV>Qќ"m&Z$KmJxӣRmnxtk` APP;{d&[ F'T2)#4Eq!x,XGE QŸ::N lrއis**EELyP$ȺDibɦ`̹C^C>,^q0t?/@qia@hń%㡮.e ؖat[:R 9 yS;ˎsN %dMSvד_ KW[R\ڄ%ʖ 9dƂl}ŸCEEF{nC./F"|_d/auVW2"ފx˂)9 B3G_o~d10.S5~ů>>O8#>4?S38B7Wz2AwjuGKoY(2)h0qPRYY[=P q4\q |V!l) l-fhH+nzaXm+UoKyLE˿^NV:wttt}kC߆TP lL1!YP ̨_o^2-c g͟9dEu Z? V %G'R.׾x_֮URCKo2'h;]NZgLoGQj(d$.[@pAP{acsV0ET"t\g5bD"*2teAE[;S ZRY*DaFBВ̾`:rB)5%QD@MBˀbper޵t[EvGbQ2ή(JqaҙgzȌt$z䛭DmTL7g#oq.Q{MՅ+3j =e3+-rы?:'h1|#ݼtkJS Kk#r]zD]qz}Q|[t%o9 "hld02K|xh_.tL.O;ǶK*^yʷfy1z/Xռ2IwLuk[wg8?D w="kM#U~G'e܃&nX5[V &.~k*⧚JmS(|-cMju{WWJ+b]y$rKq}7KZcx:{;3Q!Bg`KTrX/iisB4,>n@Kg`T-v1eԼvjZ!Q1sBF,warؖ9uQk`uPMȏF-ٽ3s=~]U ||~ӣU(?hP|^c\/Yz=/[g}Ү#A@wv <)i }`"ƃ4ͯBwi|qd b,6C7D<)IC޸v)uk= `ݺFuu{>TѽY֭ y*I=xu3Cn]yP:]ƺ=ga0{n+[UtMDNT[s?躽A ̐ \<33#8Y)%H[OXdh.kF+R*uKp#);irbwq )g,%m?VL 8-|UEV'-)qNZJ0L,]k6[ꨲ%թJCVoaB2aUT RnƁ̯rˣf Y'%VԲ+y\@O=8 $wYHt # 6h CĞTl806'2\IwaiWHrPO(.Ǜv> J§Ȏt)w ` Q[f٭(Whk*-CAi0|F.z>Xx{cP:W5Xv!UrgHR: ]`(eB(cV6sU늬,t i[$9mTn&ҩYUc~P}Њɜ6;>|NLm1&ߥQU^@HЍk=R_-o/ʢNnƃg8ѡk(l$3oY9F7\^zhBrI0V%ǨG߇6dL&w п >'RPԥ25@F0qLDѵ.Y-fa(azokɵrKw>:t%)h< Tb0S4Ơ1Ý&+M*^tdj Ud `IuZZ>7SVGmf7+݌9S '_+cryٔ{%V m1zޭmz|?3w+D`"/u_v-tNhxvuJrBٞ"1L"kv;sDsvrAٞKR7`NY/OW8 WW8- z̑g.\` |ScfUAӠ;rSry\&0?Ǔ tAOHK<2YgJd;nh m/kz$K.dd5c㮜J6RM_bJW~0J- 2rw7jtf(en)s1 9*2q JHN]̜ /YT|EHDWJ_C/ۻUX+C7D!Ę b`he&!Z0zm%QfwнN6gRwϥ9gOC-Ը/M2uIfkIgJRJ)hr2g^hbJ3#[O"G4YHZ*PfAgEfƊli pI8ЦfO24,$1%dqQ /Wד^O gRj>5hjF gl+/ M,~ wKVI,edg b !!MaCcϦFG)XR#\ʽڡps|/MhU%jH5oނ(ӻJoLm3oZ2G^ӷ"]0V.-xAoG#) Ej]$VFJr_3*Jw-u3ZRf^=鲶vš{`3Zc/]^=) hA"Hͣ \dIe9P/|×09yI M8Pwo5$!SAb쟃KǺ)[{":L87NyQ!rdhĐ;d5YUJ&/le[CA5=j_p8u9O$ɄnJm@0A|*ݘ(CHF\.9#1Ok dW3>Yubh_R"0hwEgr } e^8W b93KZ1q̺Tq0bȵo8oC%mYz6G}LS7yf&UTqjμt]6gI{8+$u|ʀD8Ȯy!TUWK)=fH"!=R]zʄZ̀1\plKs3v"xVZ=:Ҷ( 0Uk{BK }ij[egg+;Q{ aoR)؞QYŀ*ڪA`E,Pe}iԢa+5),HWK5Zxn5[3r>aO$1}~qYGuᑢX㲎?$?g"g,M{76"`1cBV'^sd2g\5bڀGeb- :=)VP yリe|ƗQ &nAN"}Gޣyz _˨'\]ݱ&#bԵ B`GURA*W kiP= B"鏜WdX6Ȉ.b!33&5GJI#9'7T(tqH'khL7BnuPt~fclύ!\+p?fp9C1,3L&i#ZWvGrՃZ·~"}*(q'3\v?}R-]ʁ~lqȴ{'c <$Poq7 KG~~sJv*WFN~Sn3CZ'B3)!OO.x}b&9 ÍٙJo\./] mKB-iOOM?sIܭT^c4䮉׬Ed[1+oW,m.f-sut}g9}!yA+{|u"f1I=N6dFw|pIN>;.B!z8w!q4%=CH'=,CHˈ#Y:zoytrASRHA\js[dGo\!a^yL,sLD^ˮu'rJԬUtlY+&ιZ-[b9GKJM@Ush~۸-X/3z]L b= d@Lr֑RCRBmU;ipΌz:yÎ@gv}ePce70Fwq EMk݁M j*U}Ӄ,| fܷdO m@[D b@L`ӈo~pD֬BgB/ю-eDVQxis3kerP[2֭MeEP -bثfBQpAhx ǀ RTKՂ #0AR3{ʥܴ֦ʥ1o +irG#7%rA%aa{xu3! Uq=t$TYlh_;/f dF:>(lDi .!%#iebYtaj`>E5j4#O=ډT/f WTR2|ϩx09]ٖe.~DC9D;&$ #([Fꀳ8JJâjyj+VVRDhY~.ҵfכËTbcp[`r]BNhGقꐲ8N.S,eՎf7dz 3Yϫ;r稗ҿ<3ƍ hx綵פČdzA`ZP 4I @i.t5pT-E$L![>rYkrh\GojчLnTQ0]m s:!lf2<8aچd12xj푡״3k##xy>e (ej- MTVZ9r_  )QBIwlMshM?zPI knGӀT( $"C8Cii_HkQ>d7'L%̅(D,wX57srX<~8n|Ǩ Qbꙉ^-g?u® G ( rC'iKh`FR6eZa'T6O&~ ,Q6s6) !D6R.qP>&FA8Ѵ/is=HC$VmCoȍJع[g9:Om}MЊYw/dwpGQh́5Mu?K{@ɾb|zTMI\B"dOi5SҲ;wwH},CW^P !E6tL! [%(Ĵv웣5h:Bݻ;ԘJ&}Sܙ~lx5gE7ݵjz=.ɽKĠ6.CHv]އت:DjQHֱ{tp$'7bDjJ(HfTɛKpJdzDjgN.*QTTÁɸ8V"K=&R;zApDꤶzLX&H-v Li:K6*Ķ^n>tyE4m;f zqCwwdgIB`G҄ {Mwlv#Py%"[.5о[aWwjJV}O&[RKF`[ח)Ӣ%m?:)}X.Iw6YB┴8%0=KCq yJL+ڷ-iMAymsFj3@p VPl|lgxo JBs{ ow}̯>\A?=[&WWd8,o{kc ߸JsB_occR$>qj#N>L]XJ]=zSMclLpBTN'AU;P5[ 9ۺal$pi'sQ7GrM'h)ā!< >c{76k"2^̑ɜyA)j8`ę`ƺ[.L@Wo!>"ɋA{zWłc,hڙW2(j F5&2 sQj)%rꔢzݛX֞ryfOCnj.,;3{̈*SJՌn]zYb"JLV20+wV[%>-{; ʎ\W#=0@n >B6ظV5Ͻ7&zfc 2W2_V+lkѫ۰_^̐I2U M@Sy\D-FnY}Jnpo-U{cQ*(C;rF*A=b^h=Q9tdqs-.=2{oӐ d{EwZ΄bwZAf倣2]-ɑÜ*kBc-x 7и:4HIE;D^)#"-lm1/%^E2HFa=}r㟁+1KH~8Wϭ~= NƄƉ(-N`);NQ/@xp[a-sLҢS2u|* Rd̐( >r]R ݢ]mAѥw':Պ6~l|,~Xbx͞*۴Nm8WߦvHOuWy Zpނp9#)@ 7EGʮ:oZ8Պ,g1rr!:ѿ2yV;䆣a8F҂wϰh`uGu[`7(>/{~iO{~ٽmv/mF\vcnv8Iw;^)%&X6buiGcaFֶ ܟZ}~S\ &z6*B`ېZس?BkJApbuF.Rmtţ9bpJ $?PJsí)&=L?Я{ Kߝb]N‹%(3ZFNTV:Lm2wHI&Iu|ѐvX0ԡpQ{v,p㑤(wrv5?F, F1{p6eLZd{Graٽi+n>\Vxݜތ>XYv:) Yn=S/sJn(_Y5֌u>Eu)[9F-D~PJH}I] 9gf́(<_ScL;7?c{ew?k&~< t̺}hrGJǟ5KžnPp;^:683j8Ҟh`'fu6.n{" X&z2ץF2"|G(h( ][ }ѠDC_%Gѕo:V7rHc4&&el0UDtL+s1gW JCH`:LFw6`J1~ߝ58Fh'0#=`( нm'mѼNЉlD~}0 {v,=!$hc *Õ4ӬͮN6 >H9@/3.%=}P7Jes9fe~0*h稳ZZ;z9AbOq71t{XC6>Geȵcan51ށ;'^޽Ħw}YczclQ7HRƆk_[؉Ԍ#]z]Y\.!7R~iE{p;FØ}=OT$=)4~b+WepJ:Jq:}SuMևzOIobzpR':lGfqybOb5Xz~|1qA+ oҳtY/唅r561-o[KTݻ{q˅ςbPg 2m28qO_OQdmyݽu5ނE{6<ډG;7F˂(шpEj<N:Yzy?s!oڃdCQoND=Q~7\|X,ng06`GҽCZN h1'.xF*³B]J@!iu*eNj-Jyv}G6QNoI5nvhPF!H$VI+`2B r#yxJ' E/Snj &A#a8Z~>m=PQv58uЦ[ Dl?;},nMώ%Y܋=C.>^qT߃rGD+ږ}0vCGNu#*z,1{r= swpm"f+C;CrqnAhiNKˑ@`Վ% xwdXC ?ۆ McM2ʬ 5JI2XkȳfɕL&裴^XG<m𺱼EF)52//O׺þ޳N;bYL_"=t%7,E8#{jXCI b5R¬,7 >E;ÔOc(d]nj!o _6?[jP!$e:}t1+9A1 dm$7{] _*cT$tq)Cbd_qHH^+Hω@dH14'[ EѝnNj\m{jײ>?/`f曳.q]^sdʊ=|RU5Y%Qsk`k⡔AeByUKfcFn Ӛ6pE蘊 6]9a(v:kvB]#(~?,.?ޭV`-Oɚ\ȸzk,9d\FZ3j 6>§'FLt?>;೬ɄZ,,+Ƴ'G?G|Qy-tv6v MHgőF{DCJFFh6^-2ȫjt4i/ӧ/3-JAx"=8&Fg$'YC)e4=t<ˏ~ F$kc82\&d8 d^zO*ɰN`,Mo,\6w3M&dHI@62i߹ q9n96:3 GPhu.>u\p*je(t#3UK"0,b騦gFFn"|d!8u ˟T^ ɜ'B<3Q0@`\&1hH$zGl(Z+ڐ*{EJ;, 2J3*(NJ(B9C.g_56c42#b9J%16Y8\gYDzz` y#DżXδ%2cdOAu: ֞|(l Pr%KYЬ8W/- Beu)roA4Sdz,dq'v#/8~\̮~ g{U4:%Y\2WSyzk(Zr_("2T-MVF Jy7?聞\ɑ,vX5ۀ|Մ; gELG0PXqVq ʴǴۥ\tOTZ%iiQ${%|3#E VrC ֐aid-Nǡ؁:,``F>!R5ƹF=23Z{r6_Ǝwa |d%+@IE1ҏI,8ЃpL)':F1`NYDz՝!pNHwFgU/=Vc2"ZRU҈Ѩ$j#2gc=bRhAルv%8kVp`gCSDQdI{՝P DtlW&.ڰRf-d>80~" HTPXvߴ&057G)A^P>7Kzh<³cpJV b(8)6'+uU/$wat*gA0SY=5Mn {kr^xAvטd.,&h: )!th4V +phMKvv9n2e_ezh+t4/JpêtX+T+XݘÂJ0:hZpp{$5bn 8y:OKET〥jˢ%VF3ZV7U2 uhKd PЌ3MxoV 넔gQl҆h,189}}#{69dA(=zJReH!:ZZ:H9 /b$c —|g1(e+oQ !7)Y\C/-O], 3c%[ (0/14Kx<:Q!Hel:Ty 1ٌRKLl %F !<ՂVW*1MX7^r8HkrR$=-MY;UhfQdzC UL7ck5*἗(% ;/! hA] +(s!{AEu~R0%[P0^HpvWO'A-i7#&Km\VmM1m;/l'E>aOڵPt(\.XVhG_-ܬXԲom4 /M %^5hmkAyÜԍ%i9B=PV-gM2 bG0q+,ïċp!,;o^MA\@𡱵Mc?>ߟe#'}cԻ1ht1^)szËE Ea0p?˹R߲e*ӿW8Ύ&,#7x]&S"R$t)[JBO m yty3«_zS},ԫk4WL!ٌk4$Mx&u: I7yF0묬iem4-|r^1q;L$`MJA2`Uo18OF4PV4(T“eL!x4yW1VMR|f:묌tq3F=Ih!暁v vZ`GT1?N.ͳW\-|2PI8{Q!PJC`Afw4*nPklȒz\% K$pl[>7z77}j r9H?&3$P=,b!cz:>T~:h69–~en敋.SGTRÃ`ym*+,#R;GxT\Ã?\weM}+tky }$]ᱛ;xg.U7/ߖ+@rNnp2Bj*>phꁣuDNyjN_86*X/X`MyJ<]Jf*][ΤtA,W}wrםo]JcxT$z.O+i00u5(r@S454w_@b 7s^E@b-JʖImdG-ҘZXյ]T6Y^wI%K/[mK LpVp?3`A"AFa1,2T[)&`%bvQ5ϤY+em3Fos,f' 8Iҩ&J< ͽ2"2BW ׊_M{4Q$l*MD!OlE;ۀ%Dm#H]7vn6M.~ٻճmh֒EcsyBO#~=3z\$JMq gEK∢{M1kw~mz׻W6컖+0xcPZ4Ÿ>Wrͯ+J㟗KmRQgJ-Om~.0G_ ˿=]~?%~sIƁoO^\mImov3{m]'շ/IwRTL@((t+Ӳv\`| djүG&yгWIp$^4HOxCRuER DyeY6ZJ+dk2*Q$)ыj mp7?|$I\9 o[tonQ VSZKL|8OnB\L'5x V biENTL(f`bLZݡm0X=1x=&̻|PtT@t+ \'gpȍ Y0?G2j%% cXO/e͒S XʦU!(ʔ2xi@ǔ t 5 릷[[a}] U7 BwX.F9ʖ9⻿R#16xLNnF*KN 4 8LC(f"7^kv n!qWi|/4Xm&fq=œ4Le=F4j 4C]Yv/\s|ݭÇknfrshgo ;m[nr {dVv37r8Ru1mOG/X?N2m76w-Ǻdq#ܫcE^sϵ'B#eD`׳ ?;gRQp>Lh2B0hm54r$ ͵J9"V-=%FR[=>B(de8 кuP8>O3UY-SG?؞m ӻ^2{T[iFi[v7H-ӖZ*3ջEF#urgʚq_qe}أ;gvC#rI:6/(,LeJlw=eH'סA5:LИ,H* t#ɉK1F RKGinL&hrX.ݰJb X\baof CdB>Ml &)4QRDeh̬&H)SRJYEZQr*{"9HPeAk_x?D:C$,b1ڽx<*l( <\Jے*D2͔&ɡ҅s>ȑ+]w@ 8uwp-L 5Bm0Uc3ݦ%!$۱4JΜA=4Vĭ Ķ{㵲Yt\ k]_ik2d\`F~ȓI 4 YXo\fF5*"M7j$ٿ;:ͱ, ׳;B&KoRB0YΓǏ ze<y ,$9e PB4mFAq^@&Q❑6L?,;pV}quK'׳!e%H-OɆ|/r;'9XYۿsev:I3_N/Kwq郤G9;ӵ?'O'R ޜDY d4SuGduF֔s?μǣKbLNF+EAՒ u9r ެksJ*y ~AԊ4}uWƆPt +/ \$ӇrefҔ^Fn> Ӊv2mzڗύ?}c '0Cʨy.Hw @p6J%+IˢKe7?Ṇ֯qtBIsD9xg;8,b+Pǽ_0?ɍM6/n@ 8i`kթvxXwDDes"9ng^jY.֧˜#UۮIxNu ղSÏUv{S՜~ ~S'iYgsmnYs-޹ד1P}.5V>ko|=^p幝KqȲ\Y^+7Cvmn ^uo~aGTn ^N[n,m qn,TR> B xooN51Վ}7jr5h_&p7ԉIS-z(.(O.[GeWǫ(xzJ#o/Sя/2xj_74M%VhºWlsn&\17?UopwV30OYK4䍫h/uS\>l"֭'T;:YJn+kUNaz{aFuu{(d nlUZ!4䍫:e9LJl=+:?UfG*cef\v/.aeB\Km|MU9幑ФGnʈy{'V]Nߠ{ڟWZ Zqܷe҈;P$.^/%k1k%/%ܦ̔!8p5D_t΍rŒre+S$/͑hG?)u}S-2Dbۼ9" %僭=r,&g>6(~d5?}~e΢I `}qD/_]kr8\9}DʔrvVr53g42'g{q]umU |f2Bt\/]6սA8Mq{FT,hSIĹO4Vga}ҞcRIYoBPB%FQz.LL[r%wѢn =b-ezuްyr>CḪӗ_ /ʰؔBI*\AE'9Clz,P٫IMpzȰh4 WR ɷu>S: ~uƌˆh8Kfh6mO>CP*Z"uN1ޕA㕕xHHK PdZ )hARLf#E19^7dm5ۅ|[Ȑu&ۛ3ZNy}H2]A&ϔF:*q2J_Rv^&GsidVbr:K Z$AD),xJL O-gו#\)חY( y~`NJK6S$~ 8*նOص{½nnWa-UyP\셖!UhPde Tnt"pH>Ā2fH΁+Nh/ p ;R>e12L5#-K/!`.(=?lm:k.wM+)hv5݌P|QHYaLiER*\unG@l X 1O&N},"FY L$sj\kVU^hTUʎxq7v:K4W>=`QlҰ:YNX%rDjelL+[JGP!īSA$_T >;u@IZ Q{˾e2W|?`U_Ь@^B+/<')D|лJLnDr-Rf.n,h(OsͲ#uhO!?Vnl~?&$er(&A\z%ށ^*7bWď1 ̾ƥ` !ɞ9=PZWx׍?#e"0gbA)Lekw$7{dwZ~e-$tMImjT:Ϫ<_^eCN9,2@ & 3m8.}$w FN8#oaD6Wv!_`ĻFVgIJl{;ozV[{[gq̓I`"2cd`cd`VlBQlC]@J|sr5us%hwrp-|#r5X.\E%TQ:ŀ *!5kmˮFI_iJ^eˮ НcAE3؇([e 7kTc+1w`$[nwQŰtXKt~r/K@Z,@̣FK~Gq[/u=[Rk^6NP?<\\Ͼt:NkCN ѓ:˃[8;9nb[J`Up<,߭S'I(eys]ZF(P:n_륡.VH ξh ˆlv<5ڴRUFa}ZqHe\@tF࢏)d ltkӕ7@gu&jkЛ@6H67C&64 $Ф 9"!&:;zi *( x<đj| `cŲ)((M:|k` Kr`L.tA<ZpХ͘3wV,5fƟ>:t Ҭtza+FraZ0 U#_e%ł}Z4(luzq$|q3fv}3/oFpIQyil>|,nŢ员vxͿVVJRP*Vd%w#+AVRVdѴ_ndJ%)=Z >/~2o$h 4NXZ9kGyruqOIhJ2}7m .GVoQ?{mah|? lM_|Wr%9Kj$Ikr8wE9C}FIw=e1[o\Ob_;y=}HE1(ـ7M,|A%w>u! M1ݥ b#I4"dʠvV[<އaLo6|#%ͬ_ew#`j~}uU|kfzoQ ;&Bz3_3d#8n2ng|$oE,` wd&iԦ{SMޟ^?=[(wLCiуs8^9QQ^2\o4uMmܮ7/҅? K-jo| ^gc B<-x6fHSL$x]yI/՟s\M&Q罬|/ZDKaITsD#ݨ.+Z G 2JO{ ߽7?.Q 7Ų\K ;{|?u8 Wus\ `6nv+υKn}Hg.LRagSѹ}PWq5rQ.6]Z $\V1B%Ux ,X,i$4 KOKWSLS8s3 cDR0 F $wl(ď/o}zx I?x:Tm\#z*;e,1)h½EXn}fQqY}ds)'PGZ#oEa/{#҃Ȉ&@*j^䵗5C0d:m|shF);Z*l QizM'ސO.S"l 0 J$x?*E&S|hk@&]MŎA,̙c;ɺU(E$V9 㦂R^ؿR \~v5Fk?k͉FB3,UY㔃@Sэ0_(ٖ#"b )uӅS!#KLaMKc倉BALNjFoT& ipjPCF/BYBzOtd#  ;NKqㄩbCn2_ɺZ,bF&'v*N5 3P=G 1ܑ`H0PF63y]sYnTh@VW 3U! :1$lNyfM/)]y &qn<*Zܵw$M*1E(j4W4Tl7)ܾjq']$NQL>&Sͷp6\\}ƯMgU{/_'5V!$ĩ4{#i4ד?fnSj[jSm[럕q߯VWW)u{T !Gv5G\o+~Gu{\n{[2dC-GRB1mx>r\ch0Ix{- ,J<L9>e7) Qa#C#4R ,垽67K?*Rʜ b ;V*Ry0pכEJ޳g*M~3fydg* X"5RbSKD׌sPNHQw,::d{>5MEGU> dg f]_/AS¸hztV:AuUu_d ,@+h4E5s~A8ÐECԕ 1[+m5@.Cwn6";653L~8h&grrA%銅cMhB&BN13x2Q[gp<)zqA“|6%X $=I?rQhF.u[bM&R?X8+w=TRmƘ9dPjvV i BC RadYIPK^( r$㦐9HP>]ZA|ItX2G'|t["Yk1 ٔ2^K\e&6Ә2#s#P-AR\g$q8(՜0 1ٸvp$"*#Ρ$h28T.e`$zk+#ݾGj턅V X0d*lьBR^סK %tH<ģGљ+#o$Dp;DJm%|˱2]5l z_sswpv &kA_NiLwZi\sou_*V>o| y f1WaBj#t HmVRLqbeM!V0FQ-1B, ^|5pCYu}B鋮'Vy bf_ޒ(p Qۑ ~3s8 e5@,f}nF]OkEdv_Vw/&tc~d"huaZ. wa8)0.OWIBu\[/oHITkK-pJAFk0jI kC ja N.px_({slIL8B/ʩ-&N (,h1$̷ C`hd &'479iq485٥!d<@2bH@v-<z稊UN6Dע=Dd.ug-,48Z03ZQ 9%Z[p9?2y0;\u]bw]Z\'=W>p"əC^,FxHljsS9%饜U몜a0 `7᜵ApZa^:Zib$> &a6M9i1`Ebx7G )1Rq`L!ҒW 8F cz#ILϠ1adZIi 8 5QJQ0K55gZS!RadS)҉VjY>ASj}OĐOc5B< ʢZPI1 Cj)%P@Vy%0R^HH{| Y}淛B( V7:$n" 6'ѻ뿵?}  n;7t= D3[fBYo XP۾ Ϝo_e\m㺲б͔ Bݷ'PDH!FN-W*y$9xJ$$%~o& =g#~<`\o3@)IQS$r"3 [::_S]Ӯo`l0Dh!.kh < 3$OVOf =f%T젔^Gc\n1!,9^x^r8_k3>&`I%`qUN^^. qqW@1 "gnj MH (?wWdB '= CO{QD E5*T{Gx.vvSFFX͡ H{~nz¤,GtC;:ihufPMLtjMCMbkʳwdSwCo),6o7uj]%ºmGmOvO-O_6{Rp?(}Sًp;>ru1[JirDjo {W۟K`#=d[El<=Ig.%2E)ڍG#jR RD9mn2JSb$:n6K8L^ K"b3(F`J?Qa㪮PZdLPkK pfDS~MDx3l8DdfDZqE%ᴮîrz˗NM4PلuiTAGx$huj^y o(㋴C`$@ !}I'X@%–VxUL.9G56`8Qc."uJ( V# P((1cfm $N\y[Ŝc9I PÚKmhm5LiN;B@ Hb8rDA$SA_B30}*ϵ1km7m/zCC u19o8IRیyC  eqb,X. ,pOar1jyPų+%ĩ)&vP_.5?8ZdL%?̪P+ˍ'W t*8Y37,E30:dsig_ eaw8IsoO6d:9ֳlrO PXh+rYt|v[_W\r\`X7o=E?]"n 2D,SW8-zɌrpV޽:\M AtapSC,߿9d"AKѵ|eظw>O [mC+p+-|ը2'=g V{[,uP$vr"}\^g3Yw2#s -T,E).gJ!\)W>Gϳ%7ZW4,R9!k7Z?XDb`hX"Z*ĜVU`Izcd L.=i)ٚnd6](5̪k 1a*͚%OBhRIc`XHf4cjphܪr42 Ѹݍi_kvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003272232215150553520017701 0ustar rootrootFeb 28 10:35:41 crc systemd[1]: Starting Kubernetes Kubelet... Feb 28 10:35:41 crc restorecon[4696]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:41 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 28 10:35:42 crc restorecon[4696]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 28 10:35:43 crc kubenswrapper[4972]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.513593 4972 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516732 4972 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516749 4972 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516753 4972 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516758 4972 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516763 4972 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516767 4972 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516772 4972 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516785 4972 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516788 4972 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516793 4972 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516797 4972 feature_gate.go:330] unrecognized feature gate: Example Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516800 4972 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516804 4972 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516807 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516811 4972 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516815 4972 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516818 4972 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516822 4972 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516825 4972 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516828 4972 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516832 4972 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516835 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516839 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516843 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516847 4972 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516850 4972 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516854 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516857 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516861 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516865 4972 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516868 4972 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516872 4972 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516876 4972 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516880 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516883 4972 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516887 4972 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516891 4972 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516894 4972 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516898 4972 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516904 4972 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516908 4972 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516911 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516914 4972 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516918 4972 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516921 4972 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516924 4972 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516929 4972 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516933 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516937 4972 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516942 4972 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516946 4972 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516950 4972 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516954 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516958 4972 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516963 4972 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516968 4972 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516973 4972 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516978 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516983 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516987 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516992 4972 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516995 4972 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.516999 4972 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517004 4972 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517008 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517012 4972 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517017 4972 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517021 4972 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517025 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517029 4972 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.517033 4972 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517887 4972 flags.go:64] FLAG: --address="0.0.0.0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517901 4972 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517910 4972 flags.go:64] FLAG: --anonymous-auth="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517916 4972 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517922 4972 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517927 4972 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517934 4972 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517939 4972 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517944 4972 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517948 4972 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517952 4972 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517957 4972 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517963 4972 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517967 4972 flags.go:64] FLAG: --cgroup-root="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517971 4972 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517976 4972 flags.go:64] FLAG: --client-ca-file="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517980 4972 flags.go:64] FLAG: --cloud-config="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517984 4972 flags.go:64] FLAG: --cloud-provider="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517988 4972 flags.go:64] FLAG: --cluster-dns="[]" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517993 4972 flags.go:64] FLAG: --cluster-domain="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.517997 4972 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518001 4972 flags.go:64] FLAG: --config-dir="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518006 4972 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518010 4972 flags.go:64] FLAG: --container-log-max-files="5" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518016 4972 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518020 4972 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518024 4972 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518029 4972 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518033 4972 flags.go:64] FLAG: --contention-profiling="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518037 4972 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518041 4972 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518046 4972 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518051 4972 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518057 4972 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518061 4972 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518065 4972 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518069 4972 flags.go:64] FLAG: --enable-load-reader="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518074 4972 flags.go:64] FLAG: --enable-server="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518078 4972 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518084 4972 flags.go:64] FLAG: --event-burst="100" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518088 4972 flags.go:64] FLAG: --event-qps="50" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518093 4972 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518097 4972 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518101 4972 flags.go:64] FLAG: --eviction-hard="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518107 4972 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518111 4972 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518115 4972 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518119 4972 flags.go:64] FLAG: --eviction-soft="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518123 4972 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518127 4972 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518131 4972 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518135 4972 flags.go:64] FLAG: --experimental-mounter-path="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518139 4972 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518143 4972 flags.go:64] FLAG: --fail-swap-on="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518147 4972 flags.go:64] FLAG: --feature-gates="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518152 4972 flags.go:64] FLAG: --file-check-frequency="20s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518156 4972 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518161 4972 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518165 4972 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518169 4972 flags.go:64] FLAG: --healthz-port="10248" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518173 4972 flags.go:64] FLAG: --help="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518177 4972 flags.go:64] FLAG: --hostname-override="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518181 4972 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518186 4972 flags.go:64] FLAG: --http-check-frequency="20s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518190 4972 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518194 4972 flags.go:64] FLAG: --image-credential-provider-config="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518198 4972 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518202 4972 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518206 4972 flags.go:64] FLAG: --image-service-endpoint="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518210 4972 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518214 4972 flags.go:64] FLAG: --kube-api-burst="100" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518218 4972 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518222 4972 flags.go:64] FLAG: --kube-api-qps="50" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518227 4972 flags.go:64] FLAG: --kube-reserved="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518232 4972 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518236 4972 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518240 4972 flags.go:64] FLAG: --kubelet-cgroups="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518244 4972 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518248 4972 flags.go:64] FLAG: --lock-file="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518252 4972 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518260 4972 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518264 4972 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518270 4972 flags.go:64] FLAG: --log-json-split-stream="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518274 4972 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518278 4972 flags.go:64] FLAG: --log-text-split-stream="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518282 4972 flags.go:64] FLAG: --logging-format="text" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518286 4972 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518290 4972 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518294 4972 flags.go:64] FLAG: --manifest-url="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518298 4972 flags.go:64] FLAG: --manifest-url-header="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518304 4972 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518308 4972 flags.go:64] FLAG: --max-open-files="1000000" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518313 4972 flags.go:64] FLAG: --max-pods="110" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518317 4972 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518321 4972 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518326 4972 flags.go:64] FLAG: --memory-manager-policy="None" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518330 4972 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518335 4972 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518339 4972 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518343 4972 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518353 4972 flags.go:64] FLAG: --node-status-max-images="50" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518357 4972 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518361 4972 flags.go:64] FLAG: --oom-score-adj="-999" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518366 4972 flags.go:64] FLAG: --pod-cidr="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518371 4972 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518379 4972 flags.go:64] FLAG: --pod-manifest-path="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518384 4972 flags.go:64] FLAG: --pod-max-pids="-1" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518389 4972 flags.go:64] FLAG: --pods-per-core="0" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518393 4972 flags.go:64] FLAG: --port="10250" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518398 4972 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518403 4972 flags.go:64] FLAG: --provider-id="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518407 4972 flags.go:64] FLAG: --qos-reserved="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518414 4972 flags.go:64] FLAG: --read-only-port="10255" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518418 4972 flags.go:64] FLAG: --register-node="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518422 4972 flags.go:64] FLAG: --register-schedulable="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518426 4972 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518433 4972 flags.go:64] FLAG: --registry-burst="10" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518437 4972 flags.go:64] FLAG: --registry-qps="5" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518441 4972 flags.go:64] FLAG: --reserved-cpus="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518445 4972 flags.go:64] FLAG: --reserved-memory="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518451 4972 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518469 4972 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518474 4972 flags.go:64] FLAG: --rotate-certificates="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518478 4972 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518482 4972 flags.go:64] FLAG: --runonce="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518486 4972 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518490 4972 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518495 4972 flags.go:64] FLAG: --seccomp-default="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518499 4972 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518508 4972 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518512 4972 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518517 4972 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518521 4972 flags.go:64] FLAG: --storage-driver-password="root" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518525 4972 flags.go:64] FLAG: --storage-driver-secure="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518529 4972 flags.go:64] FLAG: --storage-driver-table="stats" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518533 4972 flags.go:64] FLAG: --storage-driver-user="root" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518538 4972 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518543 4972 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518548 4972 flags.go:64] FLAG: --system-cgroups="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518551 4972 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518559 4972 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518563 4972 flags.go:64] FLAG: --tls-cert-file="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518567 4972 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518571 4972 flags.go:64] FLAG: --tls-min-version="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518577 4972 flags.go:64] FLAG: --tls-private-key-file="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518582 4972 flags.go:64] FLAG: --topology-manager-policy="none" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518586 4972 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518590 4972 flags.go:64] FLAG: --topology-manager-scope="container" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518594 4972 flags.go:64] FLAG: --v="2" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518600 4972 flags.go:64] FLAG: --version="false" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518607 4972 flags.go:64] FLAG: --vmodule="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518612 4972 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.518616 4972 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518738 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518743 4972 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518747 4972 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518752 4972 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518758 4972 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518762 4972 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518765 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518769 4972 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518775 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518779 4972 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518784 4972 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518788 4972 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518792 4972 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518796 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518800 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518804 4972 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518808 4972 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518812 4972 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518820 4972 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518827 4972 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518832 4972 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518836 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518840 4972 feature_gate.go:330] unrecognized feature gate: Example Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518848 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518852 4972 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518857 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518862 4972 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518866 4972 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518872 4972 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518876 4972 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518881 4972 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518887 4972 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518893 4972 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518898 4972 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518903 4972 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518908 4972 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518912 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518917 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518922 4972 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518927 4972 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518934 4972 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518939 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518944 4972 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518948 4972 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518953 4972 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518957 4972 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518962 4972 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518966 4972 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518970 4972 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518974 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518979 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518983 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518988 4972 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518992 4972 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.518996 4972 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519004 4972 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519008 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519013 4972 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519016 4972 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519020 4972 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519023 4972 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519027 4972 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519030 4972 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519034 4972 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519039 4972 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519043 4972 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519048 4972 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519052 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519056 4972 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519060 4972 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.519063 4972 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.519069 4972 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.530852 4972 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.530889 4972 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531002 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531022 4972 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531032 4972 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531041 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531050 4972 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531059 4972 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531067 4972 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531076 4972 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531084 4972 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531092 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531100 4972 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531108 4972 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531116 4972 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531124 4972 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531135 4972 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531147 4972 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531157 4972 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531166 4972 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531176 4972 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531185 4972 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531193 4972 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531201 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531209 4972 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531217 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531225 4972 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531232 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531240 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531248 4972 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531256 4972 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531264 4972 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531271 4972 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531279 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531286 4972 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531294 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531306 4972 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531317 4972 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531325 4972 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531335 4972 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531343 4972 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531352 4972 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531360 4972 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531369 4972 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531377 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531386 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531397 4972 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531407 4972 feature_gate.go:330] unrecognized feature gate: Example Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531415 4972 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531423 4972 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531433 4972 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531444 4972 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531454 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531486 4972 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531494 4972 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531502 4972 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531510 4972 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531517 4972 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531525 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531533 4972 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531541 4972 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531549 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531556 4972 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531564 4972 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531572 4972 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531579 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531587 4972 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531595 4972 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531604 4972 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531612 4972 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531620 4972 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531628 4972 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531637 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.531649 4972 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531858 4972 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531871 4972 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531880 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531890 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531898 4972 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531906 4972 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531914 4972 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531922 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531931 4972 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531939 4972 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531947 4972 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531956 4972 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531963 4972 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531971 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531979 4972 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531986 4972 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.531994 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532001 4972 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532009 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532017 4972 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532025 4972 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532032 4972 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532043 4972 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532054 4972 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532063 4972 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532071 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532080 4972 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532090 4972 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532100 4972 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532109 4972 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532117 4972 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532125 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532133 4972 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532141 4972 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532152 4972 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532162 4972 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532172 4972 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532181 4972 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532190 4972 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532199 4972 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532208 4972 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532216 4972 feature_gate.go:330] unrecognized feature gate: Example Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532224 4972 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532232 4972 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532241 4972 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532248 4972 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532256 4972 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532265 4972 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532272 4972 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532280 4972 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532288 4972 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532296 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532303 4972 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532311 4972 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532318 4972 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532326 4972 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532334 4972 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532342 4972 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532349 4972 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532357 4972 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532365 4972 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532372 4972 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532380 4972 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532414 4972 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532423 4972 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532431 4972 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532438 4972 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532446 4972 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532453 4972 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532484 4972 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.532493 4972 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.532505 4972 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.533666 4972 server.go:940] "Client rotation is on, will bootstrap in background" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.538346 4972 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.542991 4972 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.543113 4972 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.547865 4972 server.go:997] "Starting client certificate rotation" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.547907 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.548168 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.571742 4972 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.574835 4972 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.576429 4972 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.600524 4972 log.go:25] "Validated CRI v1 runtime API" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.630770 4972 log.go:25] "Validated CRI v1 image API" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.634311 4972 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.640214 4972 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-28-10-31-30-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.640255 4972 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.659351 4972 manager.go:217] Machine: {Timestamp:2026-02-28 10:35:43.654177641 +0000 UTC m=+0.566185379 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f60135d2-79a9-4dc6-a7c5-620d42b3cbaf BootID:7f06113a-3928-48fe-802c-ecedbc08cbc6 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:84:87:06 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:84:87:06 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:58:bc:57 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f2:c2:18 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4e:e8:cd Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e7:e0:3c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8e:0b:57:39:24:01 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:42:9c:08:31:7d:34 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.659687 4972 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.659903 4972 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.662304 4972 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.662530 4972 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.662573 4972 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.662859 4972 topology_manager.go:138] "Creating topology manager with none policy" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.662873 4972 container_manager_linux.go:303] "Creating device plugin manager" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.663384 4972 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.663410 4972 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.664079 4972 state_mem.go:36] "Initialized new in-memory state store" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.664579 4972 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.671234 4972 kubelet.go:418] "Attempting to sync node with API server" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.671275 4972 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.671333 4972 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.671354 4972 kubelet.go:324] "Adding apiserver pod source" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.671369 4972 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.689571 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.689583 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.689670 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.689732 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.702680 4972 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.706579 4972 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.713716 4972 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727357 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727390 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727400 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727409 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727424 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727434 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727443 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727475 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727485 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727496 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727517 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.727525 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.728857 4972 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.729369 4972 server.go:1280] "Started kubelet" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.729916 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.731064 4972 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 28 10:35:43 crc systemd[1]: Started Kubernetes Kubelet. Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.731149 4972 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733047 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733116 4972 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733273 4972 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.733716 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733828 4972 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733836 4972 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.733942 4972 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.734444 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.734518 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.735580 4972 server.go:460] "Adding debug handlers to kubelet server" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.736468 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.736562 4972 factory.go:55] Registering systemd factory Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.736590 4972 factory.go:221] Registration of the systemd container factory successfully Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.736972 4972 factory.go:153] Registering CRI-O factory Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.736992 4972 factory.go:221] Registration of the crio container factory successfully Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.737046 4972 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.737073 4972 factory.go:103] Registering Raw factory Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.737112 4972 manager.go:1196] Started watching for new ooms in manager Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.737632 4972 manager.go:319] Starting recovery of all containers Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.747116 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751160 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751242 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751260 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751272 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751284 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751293 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751305 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751331 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751345 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751356 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751366 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751377 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751388 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751400 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751410 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751419 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751429 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751439 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751450 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751474 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751484 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751494 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751504 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751514 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751524 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751535 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751548 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751559 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751570 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751582 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751592 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751602 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751651 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751661 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751671 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751686 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751696 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751707 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751717 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751728 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751739 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751749 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751760 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751771 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751781 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751808 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751819 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751829 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751840 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751850 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751861 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751870 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751886 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751899 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751910 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751920 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751931 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751942 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751952 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.751996 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752005 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752015 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752025 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752041 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752052 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752062 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752071 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752819 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752833 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752843 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752853 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752864 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752874 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752885 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752896 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752907 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752919 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752928 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752938 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752953 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752964 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752977 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.752992 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753002 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753014 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753024 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753034 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753063 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753081 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753091 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753101 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753112 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753122 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753134 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753147 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753166 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753180 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753196 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753210 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753222 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753233 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753244 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753271 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753282 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753300 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753313 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753326 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753338 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753351 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753362 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753372 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753385 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753398 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753408 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753420 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753433 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753445 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753470 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753483 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753495 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753509 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753520 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753530 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753541 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753552 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753563 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753572 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753584 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753597 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753607 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753616 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753626 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753638 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753649 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753658 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753668 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753678 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753688 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753697 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753707 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753716 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753725 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753734 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.753745 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755212 4972 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755244 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755260 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755276 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755291 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755303 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755315 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755327 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755339 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755350 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755366 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755378 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755390 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755403 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755415 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755428 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755438 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755448 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755477 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755487 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755497 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755508 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755518 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755531 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755542 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755551 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755561 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755570 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755581 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755597 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755607 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755619 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755630 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755640 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755650 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755660 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755671 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755681 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755694 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755705 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755714 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755724 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755734 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755744 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755752 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755763 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755774 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755789 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755798 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755807 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755820 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755830 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755839 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755849 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755858 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755868 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755889 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755899 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755909 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755919 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755929 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755938 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755947 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755958 4972 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755966 4972 reconstruct.go:97] "Volume reconstruction finished" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.755973 4972 reconciler.go:26] "Reconciler: start to sync state" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.763394 4972 manager.go:324] Recovery completed Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.778803 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.781331 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.781441 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.781534 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.782393 4972 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.782430 4972 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.782479 4972 state_mem.go:36] "Initialized new in-memory state store" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.784959 4972 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.787784 4972 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.787836 4972 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.787870 4972 kubelet.go:2335] "Starting kubelet main sync loop" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.787924 4972 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 28 10:35:43 crc kubenswrapper[4972]: W0228 10:35:43.790861 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.790968 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.798222 4972 policy_none.go:49] "None policy: Start" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.799152 4972 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.799179 4972 state_mem.go:35] "Initializing new in-memory state store" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.834666 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.874782 4972 manager.go:334] "Starting Device Plugin manager" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.874843 4972 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.874859 4972 server.go:79] "Starting device plugin registration server" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.875375 4972 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.875506 4972 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.875751 4972 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.875912 4972 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.875932 4972 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.885899 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.887992 4972 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.888100 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889207 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889250 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889265 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889439 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889656 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.889722 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890407 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890471 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890490 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890618 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890826 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.890887 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892077 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892151 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892165 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892405 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892557 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892611 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892654 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892561 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892724 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892754 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892811 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.892905 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894515 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894548 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894582 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894622 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894659 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894678 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894883 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894894 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.894923 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896014 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896029 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896049 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896058 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896062 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896078 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896421 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.896491 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.897356 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.897384 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.897399 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.938694 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.957906 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.957960 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.957981 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958005 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958026 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958047 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958063 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958122 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958164 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958192 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958218 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958320 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958388 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958445 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.958539 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.975743 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.977697 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.977752 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.977764 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:43 crc kubenswrapper[4972]: I0228 10:35:43.977795 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:43 crc kubenswrapper[4972]: E0228 10:35:43.978363 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059731 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059790 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059817 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059840 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059863 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059888 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059911 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059930 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.059953 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060009 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060034 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060057 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060077 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060101 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060102 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060184 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060250 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060253 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060287 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060349 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060325 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060302 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060391 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060429 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060489 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060493 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060541 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060148 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.060585 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.179390 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.181171 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.181220 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.181230 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.181259 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:44 crc kubenswrapper[4972]: E0228 10:35:44.181770 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.230210 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.251412 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.255974 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.277353 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.279507 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-359690af08a1c127529368078353894f533fe85b11618192b272470b64a176a3 WatchSource:0}: Error finding container 359690af08a1c127529368078353894f533fe85b11618192b272470b64a176a3: Status 404 returned error can't find the container with id 359690af08a1c127529368078353894f533fe85b11618192b272470b64a176a3 Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.286398 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-8a73e64188dc50c399e9a04257db782e535ea386e4ce6e14831d0d4122ffaa0f WatchSource:0}: Error finding container 8a73e64188dc50c399e9a04257db782e535ea386e4ce6e14831d0d4122ffaa0f: Status 404 returned error can't find the container with id 8a73e64188dc50c399e9a04257db782e535ea386e4ce6e14831d0d4122ffaa0f Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.291378 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-93b807064a72209a3d8ebe0f5ee11055d71e9906f51caea6c2168d4cfcb22ac6 WatchSource:0}: Error finding container 93b807064a72209a3d8ebe0f5ee11055d71e9906f51caea6c2168d4cfcb22ac6: Status 404 returned error can't find the container with id 93b807064a72209a3d8ebe0f5ee11055d71e9906f51caea6c2168d4cfcb22ac6 Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.298583 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ea6cfc4d0fb3f38f507948a707482bd363c95e19664dd61f26080bffbbe8ad02 WatchSource:0}: Error finding container ea6cfc4d0fb3f38f507948a707482bd363c95e19664dd61f26080bffbbe8ad02: Status 404 returned error can't find the container with id ea6cfc4d0fb3f38f507948a707482bd363c95e19664dd61f26080bffbbe8ad02 Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.329524 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:44 crc kubenswrapper[4972]: E0228 10:35:44.339420 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.342366 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-111f8dd679a990fc98285266edfa51475ffd7e5f177159f210c5681c6113f443 WatchSource:0}: Error finding container 111f8dd679a990fc98285266edfa51475ffd7e5f177159f210c5681c6113f443: Status 404 returned error can't find the container with id 111f8dd679a990fc98285266edfa51475ffd7e5f177159f210c5681c6113f443 Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.582333 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.584007 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.584037 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.584048 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.584072 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:44 crc kubenswrapper[4972]: E0228 10:35:44.584479 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.730982 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:44 crc kubenswrapper[4972]: W0228 10:35:44.760182 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:44 crc kubenswrapper[4972]: E0228 10:35:44.760287 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.793700 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8a73e64188dc50c399e9a04257db782e535ea386e4ce6e14831d0d4122ffaa0f"} Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.796313 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"359690af08a1c127529368078353894f533fe85b11618192b272470b64a176a3"} Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.799348 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"111f8dd679a990fc98285266edfa51475ffd7e5f177159f210c5681c6113f443"} Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.800743 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea6cfc4d0fb3f38f507948a707482bd363c95e19664dd61f26080bffbbe8ad02"} Feb 28 10:35:44 crc kubenswrapper[4972]: I0228 10:35:44.805071 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93b807064a72209a3d8ebe0f5ee11055d71e9906f51caea6c2168d4cfcb22ac6"} Feb 28 10:35:45 crc kubenswrapper[4972]: W0228 10:35:45.082785 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.082906 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.141367 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Feb 28 10:35:45 crc kubenswrapper[4972]: W0228 10:35:45.165886 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.165990 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:45 crc kubenswrapper[4972]: W0228 10:35:45.203588 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.203756 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.384936 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.387119 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.387182 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.387197 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.387234 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.387852 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.618524 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:35:45 crc kubenswrapper[4972]: E0228 10:35:45.620241 4972 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.731555 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.809215 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747" exitCode=0 Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.809292 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.809365 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.810275 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.810319 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.810332 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.811941 4972 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250" exitCode=0 Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.812048 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.812092 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.812746 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.812971 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.813015 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.813028 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.813617 4972 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3cdbbf8f569c10ad88ed7156400d81032aa6a921c85f87aaec4ba5615a5999ce" exitCode=0 Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.813675 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3cdbbf8f569c10ad88ed7156400d81032aa6a921c85f87aaec4ba5615a5999ce"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.813737 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815186 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815131 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815527 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815591 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815822 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815700 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a5ca2f261511445c4e3ef53d43ef39ab545d21b0e2c1995db48b721ee05132ac"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815763 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815887 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.815675 4972 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a5ca2f261511445c4e3ef53d43ef39ab545d21b0e2c1995db48b721ee05132ac" exitCode=0 Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.816954 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.816988 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.817001 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820140 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3307e0d7e8d64cc54dbc0365a51ef07e663ff44b7bc74ea025929cccfa082d69"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820181 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5aee7f5092a4da0fb8403beab8a61e0250a4c6e6887e54e19999e88e76c80a3"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820199 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820215 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fc7cdead7cb07221486535a183c597d90a41bcc82bef09434fc5ac1ed2f0ddfb"} Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820224 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820950 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.820990 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:45 crc kubenswrapper[4972]: I0228 10:35:45.821002 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.730949 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Feb 28 10:35:46 crc kubenswrapper[4972]: E0228 10:35:46.742668 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.747911 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.828148 4972 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a" exitCode=0 Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.828226 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.828321 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.832092 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.832141 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.832153 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: E0228 10:35:46.835343 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.835772 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"61fa7cd89a1350f58c60b016532608ac9f7b1114221b0a015c2eb7aabe4c78b6"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.835819 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ae2f3939dc1015ca6ba7b3527c3fccee63e278705462acabf07ca1dc136a10fe"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.835838 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.835836 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"54f590f89ab33d11e0e763631209714404282d26e547cf7ac7d4ff48bbdc11db"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.836746 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.836778 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.836790 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.838244 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"31f61a774f617f465383b03b35ebcd49217508aed03bb3f41ba7d08ebce02363"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.838445 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.840566 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.840620 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.840639 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.847480 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848147 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848186 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848200 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848213 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8"} Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848730 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848768 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.848778 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.988290 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.989661 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.989718 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.989731 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:46 crc kubenswrapper[4972]: I0228 10:35:46.989767 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:46 crc kubenswrapper[4972]: E0228 10:35:46.990440 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.129493 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.855149 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dee78d4a70e28398969e21f3f9a1df39805df0efafdfa22c7fbf4b7d1245be6d"} Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.855270 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.856381 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.856444 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.856501 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.862639 4972 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089" exitCode=0 Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.862775 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.862830 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.862864 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.862908 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.863360 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089"} Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.863518 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.864651 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.864686 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.864695 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865275 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865330 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865358 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865407 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865462 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865417 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865521 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865549 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:47 crc kubenswrapper[4972]: I0228 10:35:47.865667 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.870936 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.870994 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.871031 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.870910 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b2931cdd11c93734a665ce4fe41e0409614dd58eed794a630bc0169ccaddcfbd"} Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.871159 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c6fec3e5d6ea3a2942df7fdc26a2abf4ab4931aef2ba4aa1cd818d603628a41"} Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.871214 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.871247 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66b24aec46ddd08f8da7f9a2e56f33f704eeea057bda264637e05fe43d125e84"} Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.872324 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.872361 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.872380 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.872951 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.873009 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.873026 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.873683 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.873729 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:48 crc kubenswrapper[4972]: I0228 10:35:48.873746 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.347529 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.438259 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.637725 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.878681 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c169fe4a9eb1ae7619edf132ee5f291127f15ee985f86587341337613dad49f"} Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.878752 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.878780 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.878779 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"146b134a383e834cdaf63b60d1a64bb03e43c8f661b684ca208917ea82851b1e"} Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.878873 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.885025 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.885910 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.885034 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.885958 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.885996 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.886019 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.887311 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.887378 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.887398 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:49 crc kubenswrapper[4972]: I0228 10:35:49.972390 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.190598 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.192781 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.192862 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.192890 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.192949 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.335530 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.346113 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.703851 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.883351 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.883404 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.883356 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885408 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885450 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885520 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885564 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885611 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885648 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885667 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885616 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:50 crc kubenswrapper[4972]: I0228 10:35:50.885760 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.886238 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.886268 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.887774 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.887853 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.887877 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.888360 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.888401 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:51 crc kubenswrapper[4972]: I0228 10:35:51.888415 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:52 crc kubenswrapper[4972]: I0228 10:35:52.438734 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:35:52 crc kubenswrapper[4972]: I0228 10:35:52.438912 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:35:53 crc kubenswrapper[4972]: I0228 10:35:53.838320 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 28 10:35:53 crc kubenswrapper[4972]: I0228 10:35:53.838708 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:53 crc kubenswrapper[4972]: I0228 10:35:53.840754 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:53 crc kubenswrapper[4972]: I0228 10:35:53.840830 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:53 crc kubenswrapper[4972]: I0228 10:35:53.840853 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:53 crc kubenswrapper[4972]: E0228 10:35:53.886039 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.133860 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.134057 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.136000 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.136049 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.136061 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:57 crc kubenswrapper[4972]: W0228 10:35:57.411157 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.411296 4972 trace.go:236] Trace[1368998449]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Feb-2026 10:35:47.409) (total time: 10001ms): Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1368998449]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:35:57.411) Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1368998449]: [10.001985788s] [10.001985788s] END Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.411366 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 28 10:35:57 crc kubenswrapper[4972]: W0228 10:35:57.559732 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.559869 4972 trace.go:236] Trace[1680392517]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Feb-2026 10:35:47.557) (total time: 10002ms): Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1680392517]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (10:35:57.559) Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1680392517]: [10.002123571s] [10.002123571s] END Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.559923 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 28 10:35:57 crc kubenswrapper[4972]: W0228 10:35:57.705325 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.705437 4972 trace.go:236] Trace[1522720967]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Feb-2026 10:35:47.703) (total time: 10001ms): Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1522720967]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:35:57.705) Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[1522720967]: [10.001460929s] [10.001460929s] END Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.705492 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 28 10:35:57 crc kubenswrapper[4972]: W0228 10:35:57.727234 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.727357 4972 trace.go:236] Trace[764139976]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Feb-2026 10:35:47.725) (total time: 10001ms): Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[764139976]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:35:57.727) Feb 28 10:35:57 crc kubenswrapper[4972]: Trace[764139976]: [10.001563361s] [10.001563361s] END Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.727381 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.732091 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.818565 4972 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.818677 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.906920 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.909092 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dee78d4a70e28398969e21f3f9a1df39805df0efafdfa22c7fbf4b7d1245be6d" exitCode=255 Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.909156 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dee78d4a70e28398969e21f3f9a1df39805df0efafdfa22c7fbf4b7d1245be6d"} Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.909405 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.910616 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.910667 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.910677 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.911659 4972 scope.go:117] "RemoveContainer" containerID="dee78d4a70e28398969e21f3f9a1df39805df0efafdfa22c7fbf4b7d1245be6d" Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.936778 4972 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:57Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.938630 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:57Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.938853 4972 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.938936 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.941618 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:57Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 28 10:35:57 crc kubenswrapper[4972]: E0228 10:35:57.943111 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:57Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.946260 4972 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 28 10:35:57 crc kubenswrapper[4972]: I0228 10:35:57.946326 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.735142 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:58Z is after 2026-02-23T05:33:13Z Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.914079 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.915737 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2"} Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.915914 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.916884 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.916921 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:58 crc kubenswrapper[4972]: I0228 10:35:58.916934 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.644351 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.736237 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:35:59Z is after 2026-02-23T05:33:13Z Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.922353 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.923843 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.927061 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" exitCode=255 Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.927134 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2"} Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.927244 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.927261 4972 scope.go:117] "RemoveContainer" containerID="dee78d4a70e28398969e21f3f9a1df39805df0efafdfa22c7fbf4b7d1245be6d" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.933543 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.933602 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.933616 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.934357 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:35:59 crc kubenswrapper[4972]: E0228 10:35:59.934663 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:35:59 crc kubenswrapper[4972]: I0228 10:35:59.938305 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.349563 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.736446 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:00Z is after 2026-02-23T05:33:13Z Feb 28 10:36:00 crc kubenswrapper[4972]: W0228 10:36:00.806180 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:00Z is after 2026-02-23T05:33:13Z Feb 28 10:36:00 crc kubenswrapper[4972]: E0228 10:36:00.806311 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.932155 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.936022 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.937505 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.937576 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.937606 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:00 crc kubenswrapper[4972]: I0228 10:36:00.938712 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:36:00 crc kubenswrapper[4972]: E0228 10:36:00.939166 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.735574 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:01Z is after 2026-02-23T05:33:13Z Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.939839 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.941356 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.941417 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.941436 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:01 crc kubenswrapper[4972]: I0228 10:36:01.942877 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:36:01 crc kubenswrapper[4972]: E0228 10:36:01.943192 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:02 crc kubenswrapper[4972]: I0228 10:36:02.440095 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:36:02 crc kubenswrapper[4972]: I0228 10:36:02.440262 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:36:02 crc kubenswrapper[4972]: I0228 10:36:02.735312 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:02Z is after 2026-02-23T05:33:13Z Feb 28 10:36:02 crc kubenswrapper[4972]: W0228 10:36:02.855407 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:02Z is after 2026-02-23T05:33:13Z Feb 28 10:36:02 crc kubenswrapper[4972]: E0228 10:36:02.855588 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:02Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:03 crc kubenswrapper[4972]: W0228 10:36:03.272615 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:03Z is after 2026-02-23T05:33:13Z Feb 28 10:36:03 crc kubenswrapper[4972]: E0228 10:36:03.272800 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:03 crc kubenswrapper[4972]: W0228 10:36:03.386973 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:03Z is after 2026-02-23T05:33:13Z Feb 28 10:36:03 crc kubenswrapper[4972]: E0228 10:36:03.387084 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:03Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.734790 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:03Z is after 2026-02-23T05:33:13Z Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.876614 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.877164 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.878869 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.878915 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.878933 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:03 crc kubenswrapper[4972]: E0228 10:36:03.886317 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.898793 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.944860 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.946652 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.946731 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:03 crc kubenswrapper[4972]: I0228 10:36:03.946751 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.339052 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.341445 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.341607 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.341630 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.341682 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:04 crc kubenswrapper[4972]: E0228 10:36:04.347235 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:04Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:36:04 crc kubenswrapper[4972]: E0228 10:36:04.351055 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:04Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 28 10:36:04 crc kubenswrapper[4972]: I0228 10:36:04.736834 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:04Z is after 2026-02-23T05:33:13Z Feb 28 10:36:05 crc kubenswrapper[4972]: I0228 10:36:05.735968 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:05Z is after 2026-02-23T05:33:13Z Feb 28 10:36:06 crc kubenswrapper[4972]: I0228 10:36:06.617512 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:36:06 crc kubenswrapper[4972]: E0228 10:36:06.624285 4972 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:06Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:06 crc kubenswrapper[4972]: I0228 10:36:06.735862 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:06Z is after 2026-02-23T05:33:13Z Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.734450 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:07Z is after 2026-02-23T05:33:13Z Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.818019 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.818393 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.820228 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.820312 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.820327 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:07 crc kubenswrapper[4972]: I0228 10:36:07.821061 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:36:07 crc kubenswrapper[4972]: E0228 10:36:07.821294 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:07 crc kubenswrapper[4972]: E0228 10:36:07.949103 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:07Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:08 crc kubenswrapper[4972]: W0228 10:36:08.670096 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:08Z is after 2026-02-23T05:33:13Z Feb 28 10:36:08 crc kubenswrapper[4972]: E0228 10:36:08.670220 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:08Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:08 crc kubenswrapper[4972]: I0228 10:36:08.737014 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:08Z is after 2026-02-23T05:33:13Z Feb 28 10:36:09 crc kubenswrapper[4972]: I0228 10:36:09.736122 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:09Z is after 2026-02-23T05:33:13Z Feb 28 10:36:10 crc kubenswrapper[4972]: I0228 10:36:10.734720 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:10Z is after 2026-02-23T05:33:13Z Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.348558 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.350866 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.350945 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.350973 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.351025 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:11 crc kubenswrapper[4972]: E0228 10:36:11.356029 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 28 10:36:11 crc kubenswrapper[4972]: E0228 10:36:11.357859 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:36:11 crc kubenswrapper[4972]: I0228 10:36:11.737038 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z Feb 28 10:36:11 crc kubenswrapper[4972]: W0228 10:36:11.821805 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z Feb 28 10:36:11 crc kubenswrapper[4972]: E0228 10:36:11.821950 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:11 crc kubenswrapper[4972]: W0228 10:36:11.851558 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z Feb 28 10:36:11 crc kubenswrapper[4972]: E0228 10:36:11.851696 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:11Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.438708 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.438839 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.438993 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.439207 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.441232 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.441332 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.441360 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.442180 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.442564 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2" gracePeriod=30 Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.733715 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:12Z is after 2026-02-23T05:33:13Z Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.976792 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.978220 4972 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2" exitCode=255 Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.978330 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2"} Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.978396 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c222f02c3999327833618ea6b1ce18bbf6a228bfa49cc6ae8e7faaccbc8d8d3e"} Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.978863 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.980664 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.980734 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:12 crc kubenswrapper[4972]: I0228 10:36:12.980759 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:13 crc kubenswrapper[4972]: I0228 10:36:13.736959 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:13Z is after 2026-02-23T05:33:13Z Feb 28 10:36:13 crc kubenswrapper[4972]: E0228 10:36:13.886487 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:14 crc kubenswrapper[4972]: I0228 10:36:14.734960 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:14Z is after 2026-02-23T05:33:13Z Feb 28 10:36:15 crc kubenswrapper[4972]: W0228 10:36:15.312816 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:15Z is after 2026-02-23T05:33:13Z Feb 28 10:36:15 crc kubenswrapper[4972]: E0228 10:36:15.312952 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:15 crc kubenswrapper[4972]: I0228 10:36:15.734424 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:15Z is after 2026-02-23T05:33:13Z Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.736129 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:16Z is after 2026-02-23T05:33:13Z Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.748453 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.748867 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.751572 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.751830 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:16 crc kubenswrapper[4972]: I0228 10:36:16.751985 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:17 crc kubenswrapper[4972]: I0228 10:36:17.736306 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:17Z is after 2026-02-23T05:33:13Z Feb 28 10:36:17 crc kubenswrapper[4972]: E0228 10:36:17.954775 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:17Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.358257 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.360505 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.360607 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.360640 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.360701 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:18 crc kubenswrapper[4972]: E0228 10:36:18.361604 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:18Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 28 10:36:18 crc kubenswrapper[4972]: E0228 10:36:18.365818 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:18Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.736676 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:18Z is after 2026-02-23T05:33:13Z Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.788823 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.790721 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.790772 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.790793 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:18 crc kubenswrapper[4972]: I0228 10:36:18.791867 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.438810 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.439188 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.440591 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.440665 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.440683 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:19 crc kubenswrapper[4972]: I0228 10:36:19.737190 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:19Z is after 2026-02-23T05:33:13Z Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.004129 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.004961 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.007482 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" exitCode=255 Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.007508 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b"} Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.007600 4972 scope.go:117] "RemoveContainer" containerID="5544be6bc2a0d31238bafb56c9e248cdec1216142a59fec98d9106fd91c819a2" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.007698 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.009134 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.009200 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.009223 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.010159 4972 scope.go:117] "RemoveContainer" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" Feb 28 10:36:20 crc kubenswrapper[4972]: E0228 10:36:20.010399 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.349170 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:20 crc kubenswrapper[4972]: I0228 10:36:20.736194 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:20Z is after 2026-02-23T05:33:13Z Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.012429 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.015311 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.016911 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.016958 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.016975 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.017750 4972 scope.go:117] "RemoveContainer" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" Feb 28 10:36:21 crc kubenswrapper[4972]: E0228 10:36:21.017995 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:21 crc kubenswrapper[4972]: I0228 10:36:21.734595 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:21Z is after 2026-02-23T05:33:13Z Feb 28 10:36:22 crc kubenswrapper[4972]: I0228 10:36:22.438888 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:36:22 crc kubenswrapper[4972]: I0228 10:36:22.439035 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:36:22 crc kubenswrapper[4972]: I0228 10:36:22.737281 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:22Z is after 2026-02-23T05:33:13Z Feb 28 10:36:23 crc kubenswrapper[4972]: I0228 10:36:23.736707 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:23Z is after 2026-02-23T05:33:13Z Feb 28 10:36:23 crc kubenswrapper[4972]: I0228 10:36:23.811405 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:36:23 crc kubenswrapper[4972]: E0228 10:36:23.815841 4972 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:23Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:23 crc kubenswrapper[4972]: E0228 10:36:23.817044 4972 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Feb 28 10:36:23 crc kubenswrapper[4972]: E0228 10:36:23.886806 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:24 crc kubenswrapper[4972]: I0228 10:36:24.736632 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:24Z is after 2026-02-23T05:33:13Z Feb 28 10:36:24 crc kubenswrapper[4972]: W0228 10:36:24.798076 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:24Z is after 2026-02-23T05:33:13Z Feb 28 10:36:24 crc kubenswrapper[4972]: E0228 10:36:24.798199 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:24 crc kubenswrapper[4972]: W0228 10:36:24.890954 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:24Z is after 2026-02-23T05:33:13Z Feb 28 10:36:24 crc kubenswrapper[4972]: E0228 10:36:24.891051 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.366297 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:25 crc kubenswrapper[4972]: E0228 10:36:25.366331 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:25Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.368386 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.368445 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.368464 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.368521 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:25 crc kubenswrapper[4972]: E0228 10:36:25.374229 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:25Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:36:25 crc kubenswrapper[4972]: W0228 10:36:25.720805 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:25Z is after 2026-02-23T05:33:13Z Feb 28 10:36:25 crc kubenswrapper[4972]: E0228 10:36:25.721003 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 28 10:36:25 crc kubenswrapper[4972]: I0228 10:36:25.733755 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:25Z is after 2026-02-23T05:33:13Z Feb 28 10:36:26 crc kubenswrapper[4972]: I0228 10:36:26.737115 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:26Z is after 2026-02-23T05:33:13Z Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.734295 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:27Z is after 2026-02-23T05:33:13Z Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.817430 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.817777 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.819419 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.819568 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.819604 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:27 crc kubenswrapper[4972]: I0228 10:36:27.820672 4972 scope.go:117] "RemoveContainer" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" Feb 28 10:36:27 crc kubenswrapper[4972]: E0228 10:36:27.820999 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:27 crc kubenswrapper[4972]: E0228 10:36:27.959969 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:27Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:28 crc kubenswrapper[4972]: I0228 10:36:28.737591 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:28Z is after 2026-02-23T05:33:13Z Feb 28 10:36:29 crc kubenswrapper[4972]: I0228 10:36:29.735582 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:29Z is after 2026-02-23T05:33:13Z Feb 28 10:36:30 crc kubenswrapper[4972]: I0228 10:36:30.732961 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:30Z is after 2026-02-23T05:33:13Z Feb 28 10:36:31 crc kubenswrapper[4972]: I0228 10:36:31.733667 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:31Z is after 2026-02-23T05:33:13Z Feb 28 10:36:32 crc kubenswrapper[4972]: E0228 10:36:32.373175 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:32Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.375365 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.377505 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.377605 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.377633 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.377675 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:32 crc kubenswrapper[4972]: E0228 10:36:32.382456 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:32Z is after 2026-02-23T05:33:13Z" node="crc" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.439757 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.439901 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:36:32 crc kubenswrapper[4972]: I0228 10:36:32.736238 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:32Z is after 2026-02-23T05:33:13Z Feb 28 10:36:33 crc kubenswrapper[4972]: I0228 10:36:33.734325 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:33Z is after 2026-02-23T05:33:13Z Feb 28 10:36:33 crc kubenswrapper[4972]: E0228 10:36:33.886977 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.572281 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.573037 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.574837 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.574892 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.574904 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:34 crc kubenswrapper[4972]: I0228 10:36:34.734938 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:34Z is after 2026-02-23T05:33:13Z Feb 28 10:36:35 crc kubenswrapper[4972]: I0228 10:36:35.735113 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:35Z is after 2026-02-23T05:33:13Z Feb 28 10:36:36 crc kubenswrapper[4972]: I0228 10:36:36.734628 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:36:36Z is after 2026-02-23T05:33:13Z Feb 28 10:36:37 crc kubenswrapper[4972]: I0228 10:36:37.733118 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.969941 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe14fe2a2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,LastTimestamp:2026-02-28 10:35:43.729328802 +0000 UTC m=+0.641336550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.978111 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.983375 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.988235 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.993875 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abea2d9a17 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.878076951 +0000 UTC m=+0.790084699,LastTimestamp:2026-02-28 10:35:43.878076951 +0000 UTC m=+0.790084699,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:37 crc kubenswrapper[4972]: E0228 10:36:37.998723 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.889234804 +0000 UTC m=+0.801242552,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.004989 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.889259665 +0000 UTC m=+0.801267403,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.013341 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.889271956 +0000 UTC m=+0.801279704,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.019739 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.890431968 +0000 UTC m=+0.802439706,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.023882 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.890480421 +0000 UTC m=+0.802488159,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.031158 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.890500112 +0000 UTC m=+0.802507860,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.036106 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.892103546 +0000 UTC m=+0.804111294,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.043070 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.892160579 +0000 UTC m=+0.804168327,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.049000 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.892172 +0000 UTC m=+0.804179748,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.053267 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.892601283 +0000 UTC m=+0.804609041,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.056414 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.892622234 +0000 UTC m=+0.804629982,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.058128 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.892664266 +0000 UTC m=+0.804672024,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.064054 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.892708758 +0000 UTC m=+0.804716526,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.069104 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.89274568 +0000 UTC m=+0.804753438,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.074000 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.892767651 +0000 UTC m=+0.804775399,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.079008 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.894539996 +0000 UTC m=+0.806547744,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.083897 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.894557127 +0000 UTC m=+0.806564875,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.088741 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46d5876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46d5876 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781591158 +0000 UTC m=+0.693598896,LastTimestamp:2026-02-28 10:35:43.894590209 +0000 UTC m=+0.806597957,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.096521 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46ac3ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46ac3ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781422079 +0000 UTC m=+0.693429817,LastTimestamp:2026-02-28 10:35:43.894643901 +0000 UTC m=+0.806651639,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.101319 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189862abe46c5bb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189862abe46c5bb7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:43.781526455 +0000 UTC m=+0.693534193,LastTimestamp:2026-02-28 10:35:43.894668683 +0000 UTC m=+0.806676421,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.106909 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac0264f685 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.284358277 +0000 UTC m=+1.196366025,LastTimestamp:2026-02-28 10:35:44.284358277 +0000 UTC m=+1.196366025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.111617 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac02d2a58d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.291546509 +0000 UTC m=+1.203554237,LastTimestamp:2026-02-28 10:35:44.291546509 +0000 UTC m=+1.203554237,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.116527 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac0331ac64 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.29777418 +0000 UTC m=+1.209781938,LastTimestamp:2026-02-28 10:35:44.29777418 +0000 UTC m=+1.209781938,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.123072 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac0391fd99 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.304086425 +0000 UTC m=+1.216094183,LastTimestamp:2026-02-28 10:35:44.304086425 +0000 UTC m=+1.216094183,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.128726 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac061bbc40 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.346668096 +0000 UTC m=+1.258675844,LastTimestamp:2026-02-28 10:35:44.346668096 +0000 UTC m=+1.258675844,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.133563 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac2489dff7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.857202679 +0000 UTC m=+1.769210407,LastTimestamp:2026-02-28 10:35:44.857202679 +0000 UTC m=+1.769210407,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.136897 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac24a2da43 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.858839619 +0000 UTC m=+1.770847357,LastTimestamp:2026-02-28 10:35:44.858839619 +0000 UTC m=+1.770847357,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.141663 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac24a348da openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.85886793 +0000 UTC m=+1.770875668,LastTimestamp:2026-02-28 10:35:44.85886793 +0000 UTC m=+1.770875668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.146496 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac24c1696d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.860842349 +0000 UTC m=+1.772850087,LastTimestamp:2026-02-28 10:35:44.860842349 +0000 UTC m=+1.772850087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.151164 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac24d320b1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.862003377 +0000 UTC m=+1.774011115,LastTimestamp:2026-02-28 10:35:44.862003377 +0000 UTC m=+1.774011115,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.155948 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac252c21f2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.867836402 +0000 UTC m=+1.779844150,LastTimestamp:2026-02-28 10:35:44.867836402 +0000 UTC m=+1.779844150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.159530 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac257ed536 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.873256246 +0000 UTC m=+1.785264024,LastTimestamp:2026-02-28 10:35:44.873256246 +0000 UTC m=+1.785264024,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.163382 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac2594f524 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.874706212 +0000 UTC m=+1.786713960,LastTimestamp:2026-02-28 10:35:44.874706212 +0000 UTC m=+1.786713960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.167368 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac25e4820f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.879919631 +0000 UTC m=+1.791927369,LastTimestamp:2026-02-28 10:35:44.879919631 +0000 UTC m=+1.791927369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.171815 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac2605f1eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.882110955 +0000 UTC m=+1.794118703,LastTimestamp:2026-02-28 10:35:44.882110955 +0000 UTC m=+1.794118703,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.175733 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac2627d950 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.88433288 +0000 UTC m=+1.796340638,LastTimestamp:2026-02-28 10:35:44.88433288 +0000 UTC m=+1.796340638,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.179297 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac392ef943 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.203566915 +0000 UTC m=+2.115574653,LastTimestamp:2026-02-28 10:35:45.203566915 +0000 UTC m=+2.115574653,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.185562 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac39f7b5eb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.216722411 +0000 UTC m=+2.128730149,LastTimestamp:2026-02-28 10:35:45.216722411 +0000 UTC m=+2.128730149,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.193803 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac3a0dbeb9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.218166457 +0000 UTC m=+2.130174195,LastTimestamp:2026-02-28 10:35:45.218166457 +0000 UTC m=+2.130174195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.200615 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac49faccf7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.485360375 +0000 UTC m=+2.397368123,LastTimestamp:2026-02-28 10:35:45.485360375 +0000 UTC m=+2.397368123,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.206731 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac4ad41fdb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.499602907 +0000 UTC m=+2.411610655,LastTimestamp:2026-02-28 10:35:45.499602907 +0000 UTC m=+2.411610655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.213098 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac4ae6fb48 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.500838728 +0000 UTC m=+2.412846506,LastTimestamp:2026-02-28 10:35:45.500838728 +0000 UTC m=+2.412846506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.218711 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac56686398 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.693868952 +0000 UTC m=+2.605876700,LastTimestamp:2026-02-28 10:35:45.693868952 +0000 UTC m=+2.605876700,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.225289 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac574a3789 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.708668809 +0000 UTC m=+2.620676587,LastTimestamp:2026-02-28 10:35:45.708668809 +0000 UTC m=+2.620676587,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.229819 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac5d77e118 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.812324632 +0000 UTC m=+2.724332400,LastTimestamp:2026-02-28 10:35:45.812324632 +0000 UTC m=+2.724332400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.235403 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac5d95631c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.81425846 +0000 UTC m=+2.726266208,LastTimestamp:2026-02-28 10:35:45.81425846 +0000 UTC m=+2.726266208,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.240831 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac5dcd7ab9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.817934521 +0000 UTC m=+2.729942269,LastTimestamp:2026-02-28 10:35:45.817934521 +0000 UTC m=+2.729942269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.245399 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac5ddb7675 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.818850933 +0000 UTC m=+2.730858671,LastTimestamp:2026-02-28 10:35:45.818850933 +0000 UTC m=+2.730858671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.249574 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac6c35c95b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.059651419 +0000 UTC m=+2.971659167,LastTimestamp:2026-02-28 10:35:46.059651419 +0000 UTC m=+2.971659167,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.253577 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac6c4253bf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.060473279 +0000 UTC m=+2.972481017,LastTimestamp:2026-02-28 10:35:46.060473279 +0000 UTC m=+2.972481017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.258010 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac6c4c0fff openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.061111295 +0000 UTC m=+2.973119033,LastTimestamp:2026-02-28 10:35:46.061111295 +0000 UTC m=+2.973119033,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.263703 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac6c4d7ca8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.061204648 +0000 UTC m=+2.973212376,LastTimestamp:2026-02-28 10:35:46.061204648 +0000 UTC m=+2.973212376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.270493 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac6d50a7a6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.078189478 +0000 UTC m=+2.990197216,LastTimestamp:2026-02-28 10:35:46.078189478 +0000 UTC m=+2.990197216,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.274757 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac6d64515e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.07947811 +0000 UTC m=+2.991485848,LastTimestamp:2026-02-28 10:35:46.07947811 +0000 UTC m=+2.991485848,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.278549 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189862ac6e638703 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.096203523 +0000 UTC m=+3.008211261,LastTimestamp:2026-02-28 10:35:46.096203523 +0000 UTC m=+3.008211261,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.282396 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac6e69d653 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.096617043 +0000 UTC m=+3.008624781,LastTimestamp:2026-02-28 10:35:46.096617043 +0000 UTC m=+3.008624781,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.287206 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac6e6a0408 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.096628744 +0000 UTC m=+3.008636482,LastTimestamp:2026-02-28 10:35:46.096628744 +0000 UTC m=+3.008636482,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.293509 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac6ebca054 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.102042708 +0000 UTC m=+3.014050446,LastTimestamp:2026-02-28 10:35:46.102042708 +0000 UTC m=+3.014050446,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.299178 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac79b1e29b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.285888155 +0000 UTC m=+3.197895893,LastTimestamp:2026-02-28 10:35:46.285888155 +0000 UTC m=+3.197895893,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.304722 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac79b40b79 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.286029689 +0000 UTC m=+3.198037427,LastTimestamp:2026-02-28 10:35:46.286029689 +0000 UTC m=+3.198037427,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.309201 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac7aa32986 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.301700486 +0000 UTC m=+3.213708224,LastTimestamp:2026-02-28 10:35:46.301700486 +0000 UTC m=+3.213708224,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.313402 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac7ab47d88 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.302836104 +0000 UTC m=+3.214843842,LastTimestamp:2026-02-28 10:35:46.302836104 +0000 UTC m=+3.214843842,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.319992 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac7abc5b05 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.303351557 +0000 UTC m=+3.215359295,LastTimestamp:2026-02-28 10:35:46.303351557 +0000 UTC m=+3.215359295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.326437 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac7acab23e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.30429139 +0000 UTC m=+3.216299128,LastTimestamp:2026-02-28 10:35:46.30429139 +0000 UTC m=+3.216299128,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.330863 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac86fb73e8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.508813288 +0000 UTC m=+3.420821026,LastTimestamp:2026-02-28 10:35:46.508813288 +0000 UTC m=+3.420821026,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.336953 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac870a9cdd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.509806813 +0000 UTC m=+3.421814541,LastTimestamp:2026-02-28 10:35:46.509806813 +0000 UTC m=+3.421814541,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.341672 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac8811b104 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.52704794 +0000 UTC m=+3.439055678,LastTimestamp:2026-02-28 10:35:46.52704794 +0000 UTC m=+3.439055678,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.348169 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac8821b1cb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.528096715 +0000 UTC m=+3.440104453,LastTimestamp:2026-02-28 10:35:46.528096715 +0000 UTC m=+3.440104453,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.352846 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189862ac883539d7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.529376727 +0000 UTC m=+3.441384465,LastTimestamp:2026-02-28 10:35:46.529376727 +0000 UTC m=+3.441384465,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.357812 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac92279680 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.696255104 +0000 UTC m=+3.608262852,LastTimestamp:2026-02-28 10:35:46.696255104 +0000 UTC m=+3.608262852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.369057 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac92cc97d2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.707068882 +0000 UTC m=+3.619076620,LastTimestamp:2026-02-28 10:35:46.707068882 +0000 UTC m=+3.619076620,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.375410 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac92dfc88f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.708326543 +0000 UTC m=+3.620334281,LastTimestamp:2026-02-28 10:35:46.708326543 +0000 UTC m=+3.620334281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.382725 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ac9a71d301 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.835337985 +0000 UTC m=+3.747345723,LastTimestamp:2026-02-28 10:35:46.835337985 +0000 UTC m=+3.747345723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.389743 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac9f6e3c22 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.918988834 +0000 UTC m=+3.830996562,LastTimestamp:2026-02-28 10:35:46.918988834 +0000 UTC m=+3.830996562,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.395143 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862aca05d6c63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.934664291 +0000 UTC m=+3.846672029,LastTimestamp:2026-02-28 10:35:46.934664291 +0000 UTC m=+3.846672029,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.399656 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862aca4b63eb4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:47.007594164 +0000 UTC m=+3.919601902,LastTimestamp:2026-02-28 10:35:47.007594164 +0000 UTC m=+3.919601902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.403873 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862aca54c007e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:47.017408638 +0000 UTC m=+3.929416376,LastTimestamp:2026-02-28 10:35:47.017408638 +0000 UTC m=+3.929416376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.411754 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862acd7f8f303 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:47.867603715 +0000 UTC m=+4.779611493,LastTimestamp:2026-02-28 10:35:47.867603715 +0000 UTC m=+4.779611493,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.418102 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ace55e39bd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.092344765 +0000 UTC m=+5.004352503,LastTimestamp:2026-02-28 10:35:48.092344765 +0000 UTC m=+5.004352503,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.423412 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ace601df95 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.103069589 +0000 UTC m=+5.015077327,LastTimestamp:2026-02-28 10:35:48.103069589 +0000 UTC m=+5.015077327,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.428564 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ace6159506 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.104361222 +0000 UTC m=+5.016368960,LastTimestamp:2026-02-28 10:35:48.104361222 +0000 UTC m=+5.016368960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.435147 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862acf3380656 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.324722262 +0000 UTC m=+5.236730010,LastTimestamp:2026-02-28 10:35:48.324722262 +0000 UTC m=+5.236730010,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.441245 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862acf3cd43a8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.334502824 +0000 UTC m=+5.246510612,LastTimestamp:2026-02-28 10:35:48.334502824 +0000 UTC m=+5.246510612,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.446517 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862acf3e9e036 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.33637791 +0000 UTC m=+5.248385688,LastTimestamp:2026-02-28 10:35:48.33637791 +0000 UTC m=+5.248385688,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.451414 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad02bf491c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.585244956 +0000 UTC m=+5.497252724,LastTimestamp:2026-02-28 10:35:48.585244956 +0000 UTC m=+5.497252724,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.456625 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad0399fce6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.59957783 +0000 UTC m=+5.511585608,LastTimestamp:2026-02-28 10:35:48.59957783 +0000 UTC m=+5.511585608,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.462256 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad03b48c53 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.601318483 +0000 UTC m=+5.513326221,LastTimestamp:2026-02-28 10:35:48.601318483 +0000 UTC m=+5.513326221,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.466857 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad13ee349f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.873532575 +0000 UTC m=+5.785540333,LastTimestamp:2026-02-28 10:35:48.873532575 +0000 UTC m=+5.785540333,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.472849 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad14df27d4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.889323476 +0000 UTC m=+5.801331234,LastTimestamp:2026-02-28 10:35:48.889323476 +0000 UTC m=+5.801331234,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.474325 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad14f1d9e2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:48.890548706 +0000 UTC m=+5.802556454,LastTimestamp:2026-02-28 10:35:48.890548706 +0000 UTC m=+5.802556454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.478615 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad2021f809 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:49.078251529 +0000 UTC m=+5.990259277,LastTimestamp:2026-02-28 10:35:49.078251529 +0000 UTC m=+5.990259277,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.479659 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189862ad20e1e1a6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:49.09082871 +0000 UTC m=+6.002836468,LastTimestamp:2026-02-28 10:35:49.09082871 +0000 UTC m=+6.002836468,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.489086 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-controller-manager-crc.189862ade870e10c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:52.438866188 +0000 UTC m=+9.350873966,LastTimestamp:2026-02-28 10:35:52.438866188 +0000 UTC m=+9.350873966,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.493757 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ade8727ea6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:52.43897207 +0000 UTC m=+9.350979848,LastTimestamp:2026-02-28 10:35:52.43897207 +0000 UTC m=+9.350979848,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.498997 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-apiserver-crc.189862af2919b2ad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.818639021 +0000 UTC m=+14.730646779,LastTimestamp:2026-02-28 10:35:57.818639021 +0000 UTC m=+14.730646779,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.505015 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862af291ad42d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.818713133 +0000 UTC m=+14.730720881,LastTimestamp:2026-02-28 10:35:57.818713133 +0000 UTC m=+14.730720881,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.510961 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189862ac92dfc88f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862ac92dfc88f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:46.708326543 +0000 UTC m=+3.620334281,LastTimestamp:2026-02-28 10:35:57.913067092 +0000 UTC m=+14.825074840,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.516053 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-apiserver-crc.189862af3044e430 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 28 10:36:38 crc kubenswrapper[4972]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 28 10:36:38 crc kubenswrapper[4972]: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.938910256 +0000 UTC m=+14.850917994,LastTimestamp:2026-02-28 10:35:57.938910256 +0000 UTC m=+14.850917994,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.521099 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862af3045ce5f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.938970207 +0000 UTC m=+14.850977945,LastTimestamp:2026-02-28 10:35:57.938970207 +0000 UTC m=+14.850977945,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.526265 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189862af3044e430\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-apiserver-crc.189862af3044e430 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 28 10:36:38 crc kubenswrapper[4972]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 28 10:36:38 crc kubenswrapper[4972]: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.938910256 +0000 UTC m=+14.850917994,LastTimestamp:2026-02-28 10:35:57.946307559 +0000 UTC m=+14.858315297,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.531796 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189862af3045ce5f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189862af3045ce5f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:57.938970207 +0000 UTC m=+14.850977945,LastTimestamp:2026-02-28 10:35:57.94634881 +0000 UTC m=+14.858356548,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.538045 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-controller-manager-crc.189862b03c9144e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440209634 +0000 UTC m=+19.352217442,LastTimestamp:2026-02-28 10:36:02.440209634 +0000 UTC m=+19.352217442,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.543571 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862b03c92f2bc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440319676 +0000 UTC m=+19.352327454,LastTimestamp:2026-02-28 10:36:02.440319676 +0000 UTC m=+19.352327454,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.551270 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862b03c9144e2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-controller-manager-crc.189862b03c9144e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440209634 +0000 UTC m=+19.352217442,LastTimestamp:2026-02-28 10:36:12.43879474 +0000 UTC m=+29.350802488,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.556255 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862b03c92f2bc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862b03c92f2bc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440319676 +0000 UTC m=+19.352327454,LastTimestamp:2026-02-28 10:36:12.438937714 +0000 UTC m=+29.350945472,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.561564 4972 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862b290c07aa8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:12.442524328 +0000 UTC m=+29.354532126,LastTimestamp:2026-02-28 10:36:12.442524328 +0000 UTC m=+29.354532126,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.566784 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862ac2594f524\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac2594f524 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:44.874706212 +0000 UTC m=+1.786713960,LastTimestamp:2026-02-28 10:36:12.56839636 +0000 UTC m=+29.480404138,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.572185 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862ac392ef943\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac392ef943 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.203566915 +0000 UTC m=+2.115574653,LastTimestamp:2026-02-28 10:36:12.856236484 +0000 UTC m=+29.768244232,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.578353 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862ac39f7b5eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862ac39f7b5eb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:35:45.216722411 +0000 UTC m=+2.128730149,LastTimestamp:2026-02-28 10:36:12.867882319 +0000 UTC m=+29.779890097,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.586014 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862b03c9144e2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-controller-manager-crc.189862b03c9144e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440209634 +0000 UTC m=+19.352217442,LastTimestamp:2026-02-28 10:36:22.439004148 +0000 UTC m=+39.351011926,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.590227 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862b03c92f2bc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189862b03c92f2bc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440319676 +0000 UTC m=+19.352327454,LastTimestamp:2026-02-28 10:36:22.43909635 +0000 UTC m=+39.351104128,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.599491 4972 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189862b03c9144e2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 28 10:36:38 crc kubenswrapper[4972]: &Event{ObjectMeta:{kube-controller-manager-crc.189862b03c9144e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 28 10:36:38 crc kubenswrapper[4972]: body: Feb 28 10:36:38 crc kubenswrapper[4972]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:36:02.440209634 +0000 UTC m=+19.352217442,LastTimestamp:2026-02-28 10:36:32.439857989 +0000 UTC m=+49.351865767,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 28 10:36:38 crc kubenswrapper[4972]: > Feb 28 10:36:38 crc kubenswrapper[4972]: W0228 10:36:38.734199 4972 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:38 crc kubenswrapper[4972]: E0228 10:36:38.734270 4972 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 28 10:36:38 crc kubenswrapper[4972]: I0228 10:36:38.735142 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.383105 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:39 crc kubenswrapper[4972]: E0228 10:36:39.383638 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.385143 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.385215 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.385241 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.385287 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:39 crc kubenswrapper[4972]: E0228 10:36:39.393803 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.736927 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.788906 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.789930 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.789997 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.790020 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:39 crc kubenswrapper[4972]: I0228 10:36:39.790862 4972 scope.go:117] "RemoveContainer" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.101147 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.102761 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5"} Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.102898 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.103671 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.103698 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.103709 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:40 crc kubenswrapper[4972]: I0228 10:36:40.735654 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.108633 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.109557 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.111954 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" exitCode=255 Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.111995 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5"} Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.112030 4972 scope.go:117] "RemoveContainer" containerID="2c5e198920248a14ae2d4f0819ab88442186395012e36641c5f2f37a4b441a6b" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.112252 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.113779 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.113840 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.113866 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.114951 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:36:41 crc kubenswrapper[4972]: E0228 10:36:41.115393 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:41 crc kubenswrapper[4972]: I0228 10:36:41.735820 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.118433 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.438519 4972 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.438749 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.438818 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.439004 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.440125 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.440166 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.440179 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.440618 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"c222f02c3999327833618ea6b1ce18bbf6a228bfa49cc6ae8e7faaccbc8d8d3e"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.440713 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://c222f02c3999327833618ea6b1ce18bbf6a228bfa49cc6ae8e7faaccbc8d8d3e" gracePeriod=30 Feb 28 10:36:42 crc kubenswrapper[4972]: I0228 10:36:42.736697 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.128004 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.129796 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.130524 4972 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c222f02c3999327833618ea6b1ce18bbf6a228bfa49cc6ae8e7faaccbc8d8d3e" exitCode=255 Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.130584 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c222f02c3999327833618ea6b1ce18bbf6a228bfa49cc6ae8e7faaccbc8d8d3e"} Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.130646 4972 scope.go:117] "RemoveContainer" containerID="a1bde4c1595111f5ae89b0a59420c3637f21b6e27a64664e0474772d0738d1d2" Feb 28 10:36:43 crc kubenswrapper[4972]: I0228 10:36:43.738275 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:43 crc kubenswrapper[4972]: E0228 10:36:43.887556 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.136840 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.138450 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bfc2d906353c13f2c150c79e6fbd990db35a6362a9850858fb24f2ea32ba0622"} Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.138567 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.139982 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.140012 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.140020 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:44 crc kubenswrapper[4972]: I0228 10:36:44.736009 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:45 crc kubenswrapper[4972]: I0228 10:36:45.141750 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:45 crc kubenswrapper[4972]: I0228 10:36:45.144217 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:45 crc kubenswrapper[4972]: I0228 10:36:45.144279 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:45 crc kubenswrapper[4972]: I0228 10:36:45.144301 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:45 crc kubenswrapper[4972]: I0228 10:36:45.737917 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:46 crc kubenswrapper[4972]: E0228 10:36:46.387608 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.394759 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.395953 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.396083 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.396174 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.396278 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:46 crc kubenswrapper[4972]: E0228 10:36:46.399502 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.735349 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.748578 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.748755 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.749908 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.750018 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:46 crc kubenswrapper[4972]: I0228 10:36:46.750035 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.734260 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.817692 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.817975 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.819347 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.819436 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.819454 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:47 crc kubenswrapper[4972]: I0228 10:36:47.820167 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:36:47 crc kubenswrapper[4972]: E0228 10:36:47.820394 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:48 crc kubenswrapper[4972]: I0228 10:36:48.735881 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:48 crc kubenswrapper[4972]: I0228 10:36:48.788760 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:48 crc kubenswrapper[4972]: I0228 10:36:48.790340 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:48 crc kubenswrapper[4972]: I0228 10:36:48.790387 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:48 crc kubenswrapper[4972]: I0228 10:36:48.790396 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.438680 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.438885 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.440014 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.440058 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.440108 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.442028 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:49 crc kubenswrapper[4972]: I0228 10:36:49.736249 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.153966 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.155234 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.155339 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.155422 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.349318 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.349648 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.351506 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.351545 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.351560 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.353153 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:36:50 crc kubenswrapper[4972]: E0228 10:36:50.353697 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:36:50 crc kubenswrapper[4972]: I0228 10:36:50.736231 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:51 crc kubenswrapper[4972]: I0228 10:36:51.735160 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:52 crc kubenswrapper[4972]: I0228 10:36:52.741457 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:53 crc kubenswrapper[4972]: E0228 10:36:53.393641 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.400514 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.402106 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.402170 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.402186 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.402223 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:36:53 crc kubenswrapper[4972]: E0228 10:36:53.408665 4972 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 28 10:36:53 crc kubenswrapper[4972]: I0228 10:36:53.736632 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:53 crc kubenswrapper[4972]: E0228 10:36:53.888104 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:36:54 crc kubenswrapper[4972]: I0228 10:36:54.735045 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:55 crc kubenswrapper[4972]: I0228 10:36:55.735728 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:55 crc kubenswrapper[4972]: I0228 10:36:55.819346 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 28 10:36:55 crc kubenswrapper[4972]: I0228 10:36:55.834394 4972 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.733247 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.754547 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.754736 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.756074 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.756134 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:36:56 crc kubenswrapper[4972]: I0228 10:36:56.756150 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:36:57 crc kubenswrapper[4972]: I0228 10:36:57.735710 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:58 crc kubenswrapper[4972]: I0228 10:36:58.736096 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:36:59 crc kubenswrapper[4972]: I0228 10:36:59.731856 4972 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.261012 4972 csr.go:261] certificate signing request csr-5s6gq is approved, waiting to be issued Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.269945 4972 csr.go:257] certificate signing request csr-5s6gq is issued Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.384889 4972 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.409556 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.411205 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.411256 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.411287 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.411510 4972 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.421089 4972 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.421381 4972 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.421409 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.433244 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.433308 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.433320 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.433380 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.433398 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:00Z","lastTransitionTime":"2026-02-28T10:37:00Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.451451 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.461957 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.462010 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.462027 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.462061 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.462079 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:00Z","lastTransitionTime":"2026-02-28T10:37:00Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.477172 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.485402 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.485797 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.485827 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.485848 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.485859 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:00Z","lastTransitionTime":"2026-02-28T10:37:00Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.503413 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.512581 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.512645 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.512664 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.512694 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.512712 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:00Z","lastTransitionTime":"2026-02-28T10:37:00Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.528961 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:00Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.529094 4972 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.529136 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:00 crc kubenswrapper[4972]: I0228 10:37:00.549263 4972 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.629228 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.730124 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.831485 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:00 crc kubenswrapper[4972]: E0228 10:37:00.932288 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.033214 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.134224 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.234541 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: I0228 10:37:01.272079 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-26 09:56:26.564723539 +0000 UTC Feb 28 10:37:01 crc kubenswrapper[4972]: I0228 10:37:01.272161 4972 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7223h19m25.292567666s for next certificate rotation Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.334666 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.435678 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.536394 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.637502 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.737680 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.838584 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:01 crc kubenswrapper[4972]: E0228 10:37:01.939373 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.040119 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.141328 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.241500 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.342319 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.443366 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.544360 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.644753 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.745855 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: I0228 10:37:02.788753 4972 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 28 10:37:02 crc kubenswrapper[4972]: I0228 10:37:02.791405 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:02 crc kubenswrapper[4972]: I0228 10:37:02.791525 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:02 crc kubenswrapper[4972]: I0228 10:37:02.791557 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:02 crc kubenswrapper[4972]: I0228 10:37:02.792698 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.792997 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.847088 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:02 crc kubenswrapper[4972]: E0228 10:37:02.948153 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.049161 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.149620 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.250427 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.351455 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.451867 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.552405 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.652600 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.753363 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.853565 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.889277 4972 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 28 10:37:03 crc kubenswrapper[4972]: E0228 10:37:03.953726 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.053951 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.154266 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.255324 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.355668 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.455995 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.557025 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.658143 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.758595 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.859270 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:04 crc kubenswrapper[4972]: E0228 10:37:04.960147 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.060334 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.160543 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.261640 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.362652 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.463431 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.564361 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.664846 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.765585 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.866645 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:05 crc kubenswrapper[4972]: E0228 10:37:05.967681 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.068624 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.169700 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.269957 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.370754 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.471336 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.572141 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.672352 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.772978 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.873164 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:06 crc kubenswrapper[4972]: E0228 10:37:06.973283 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.074417 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.174621 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.274807 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.375633 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.476444 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.576961 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.677567 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.778181 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.878901 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:07 crc kubenswrapper[4972]: E0228 10:37:07.979062 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.079283 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.180229 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.280718 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.381668 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.482436 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.583548 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.683967 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.784755 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.885107 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:08 crc kubenswrapper[4972]: E0228 10:37:08.986047 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.086857 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.187092 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.287314 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.387861 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.488438 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.588710 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.689495 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.790594 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.890744 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:09 crc kubenswrapper[4972]: E0228 10:37:09.991359 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.091940 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.192386 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.292667 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.393261 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.493853 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.594694 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.694819 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.709911 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.715345 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.715408 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.715422 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.715447 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.715489 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:10Z","lastTransitionTime":"2026-02-28T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.728763 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.738435 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.738526 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.738544 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.738569 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.738588 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:10Z","lastTransitionTime":"2026-02-28T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.755853 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.764137 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.764185 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.764196 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.764216 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.764226 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:10Z","lastTransitionTime":"2026-02-28T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.779528 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.790776 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.790853 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.790873 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.790899 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:10 crc kubenswrapper[4972]: I0228 10:37:10.790919 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:10Z","lastTransitionTime":"2026-02-28T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.804840 4972 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f06113a-3928-48fe-802c-ecedbc08cbc6\\\",\\\"systemUUID\\\":\\\"f60135d2-79a9-4dc6-a7c5-620d42b3cbaf\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.805019 4972 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.805056 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:10 crc kubenswrapper[4972]: E0228 10:37:10.906086 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.006410 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: I0228 10:37:11.080821 4972 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.106567 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.206966 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.307689 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.408275 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.509193 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.609886 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.710125 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.810820 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:11 crc kubenswrapper[4972]: E0228 10:37:11.911616 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.012150 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.112815 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.213035 4972 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.245690 4972 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.256899 4972 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.315920 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.315963 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.315976 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.315992 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.316006 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.418804 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.418857 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.418873 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.418893 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.418909 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.522194 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.522265 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.522292 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.522322 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.522344 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.626140 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.626209 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.626235 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.626267 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.626291 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.729426 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.729517 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.729535 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.729561 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.729578 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.738025 4972 apiserver.go:52] "Watching apiserver" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.743889 4972 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.744421 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.744914 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.745058 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.745264 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.745557 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.745229 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.745852 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.745851 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.746132 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.746206 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.748300 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.748419 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.748993 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.749101 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.749188 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.749650 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.751030 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.751425 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.752010 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.761707 4972 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.780421 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.801272 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.817922 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.832605 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.832677 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.832687 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.832708 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.832721 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.834012 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.834824 4972 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.845955 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.859127 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869557 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869617 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869650 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869674 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869702 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869732 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869751 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869769 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869791 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869815 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869837 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869854 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869876 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869898 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869916 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869937 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869961 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.869982 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870000 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870016 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870035 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870046 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870069 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870086 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870087 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870121 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870217 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870255 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870281 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870369 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870397 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870420 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870444 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870485 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870514 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870537 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870566 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870592 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870615 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870640 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870664 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870687 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870709 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870733 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870755 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870819 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870842 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870864 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870886 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870907 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870928 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870950 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870972 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870995 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871018 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871038 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871059 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871083 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871106 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871130 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871154 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871179 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871199 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871219 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871240 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871263 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871285 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871310 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871334 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871361 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871383 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871404 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871425 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871448 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871528 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871553 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871578 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871607 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871635 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871662 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871688 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871714 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871738 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871763 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871783 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871808 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871831 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871852 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871882 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871904 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871928 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871949 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871972 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871999 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872022 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872089 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872118 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872141 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872165 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872188 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872212 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872235 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872257 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872281 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872304 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872330 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872352 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872374 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872398 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872449 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872503 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872532 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872553 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872578 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872600 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872654 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872679 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872701 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872726 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872747 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872769 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872794 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872815 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872837 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872862 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872886 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872909 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872934 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872956 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872977 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873001 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873092 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873122 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873147 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873168 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873190 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873214 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873237 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873264 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873289 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873311 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873336 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873366 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873391 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873415 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873439 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873488 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873511 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873534 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873560 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873584 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873608 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873632 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873657 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873683 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873711 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873738 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873765 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873787 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873809 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873833 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873857 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873885 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873910 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873935 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873959 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873983 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870371 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870433 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874009 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873990 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874037 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874061 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874087 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874111 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874134 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874158 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874180 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874204 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870576 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870695 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870771 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870954 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.870986 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871039 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871136 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874073 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871243 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.871573 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872282 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872357 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872530 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872746 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872800 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872828 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.872829 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873519 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873763 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873854 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873868 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873920 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.873975 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874230 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874497 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874526 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874552 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874561 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874580 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874608 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874630 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874629 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874652 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874676 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874695 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874715 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874733 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874755 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874771 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874788 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874808 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874827 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874844 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874864 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874882 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874854 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874909 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874930 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874874 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.874994 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875027 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875039 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875050 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875079 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875101 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875122 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875141 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875163 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875186 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875280 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875300 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875320 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875340 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875362 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875407 4972 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875421 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875434 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875449 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875501 4972 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875514 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875529 4972 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875546 4972 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875564 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875577 4972 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875590 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875603 4972 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875616 4972 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875630 4972 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875645 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875659 4972 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875673 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875687 4972 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875702 4972 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875719 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875731 4972 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875743 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875754 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875764 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875776 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875792 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876358 4972 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876372 4972 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876386 4972 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876400 4972 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876448 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876491 4972 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876508 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876523 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883980 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875064 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875221 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875271 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875307 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875516 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875576 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875583 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.875779 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.876028 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877180 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877230 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.901584 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.901624 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.901354 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.902707 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.902786 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.899477 4972 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.902990 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877387 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877332 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877773 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877880 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877905 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877922 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.877947 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.878842 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.879435 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.879596 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.879908 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.880541 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.880749 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881101 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881127 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881161 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881385 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881411 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881883 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881929 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.881942 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.882504 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.882584 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.882642 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.882644 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883016 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883065 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883075 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883093 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883104 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883279 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883385 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883401 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883581 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883588 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883579 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.883939 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.884351 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.884672 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.884710 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:13.384678182 +0000 UTC m=+90.296686130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.884798 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.891681 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.892860 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.892905 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.893162 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.893640 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.893854 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.894278 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.894787 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.895226 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.895285 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.894361 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.895676 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.895705 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.895935 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896002 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896088 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896239 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896425 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896636 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896740 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.896777 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897025 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897032 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897230 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897374 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897395 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897804 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897822 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897872 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897887 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897933 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.897954 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898172 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898265 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898365 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898542 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898574 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898918 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.898964 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.899061 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.899323 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.899337 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.900420 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.900526 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.900546 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.900633 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.900626 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.901236 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.901344 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.901379 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.903094 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.903077 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.905151 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.905487 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:13.405434727 +0000 UTC m=+90.317442465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.905957 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:13.405917431 +0000 UTC m=+90.317925169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.915898 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.918727 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.918759 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.918775 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.918881 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:13.418852135 +0000 UTC m=+90.330859873 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.919149 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.919183 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.919200 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:12 crc kubenswrapper[4972]: E0228 10:37:12.919264 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:13.419245047 +0000 UTC m=+90.331252785 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.921336 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.921447 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.921739 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.921878 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.921863 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.922352 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.922577 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.922893 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.923025 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.923073 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.923146 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.923479 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.923692 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924012 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924316 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924448 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924612 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924641 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.924748 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.925276 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.926170 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.926743 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.928442 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.928736 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.930121 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.931371 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.931717 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.931913 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932128 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932157 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932175 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932256 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932266 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932300 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932382 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932529 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932583 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932695 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932753 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932760 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932761 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932895 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932911 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.932948 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.933567 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.933407 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.933983 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.934088 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.934323 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.934637 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.935601 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.935635 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.935904 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936112 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936138 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936148 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936163 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936173 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:12Z","lastTransitionTime":"2026-02-28T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.936445 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.941522 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.948065 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.954319 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.963767 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.968561 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977684 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977745 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977808 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977851 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977872 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977889 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977905 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977923 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977937 4972 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977950 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.977998 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978049 4972 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978075 4972 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978090 4972 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978105 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978140 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978159 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978173 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978187 4972 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978215 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978229 4972 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978243 4972 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978255 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978268 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978282 4972 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978295 4972 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978309 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978322 4972 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978334 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978348 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978363 4972 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978378 4972 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978392 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978406 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978507 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978521 4972 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978535 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978550 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978566 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978579 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978595 4972 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978622 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978637 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978669 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978686 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978702 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978715 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978729 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978745 4972 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978758 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978771 4972 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978784 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978797 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978810 4972 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978823 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978835 4972 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978850 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978865 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978878 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978890 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978904 4972 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978917 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978930 4972 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978943 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978956 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978970 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978985 4972 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.978999 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979012 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979026 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979039 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979053 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979065 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979079 4972 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979092 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979105 4972 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979119 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979135 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979148 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979165 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979178 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979191 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979204 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979218 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979230 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979242 4972 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979256 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979269 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979282 4972 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979296 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979308 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979323 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979336 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979349 4972 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979362 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979375 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979388 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979401 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979416 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979430 4972 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979443 4972 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979472 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979486 4972 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979499 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979512 4972 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979526 4972 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979540 4972 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979552 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979584 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979596 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979612 4972 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979625 4972 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979641 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979654 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979667 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979680 4972 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979693 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979706 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979718 4972 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979732 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979747 4972 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979759 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979774 4972 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979788 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979801 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979816 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979829 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979842 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979855 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979868 4972 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979880 4972 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979892 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979904 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979916 4972 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979929 4972 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979940 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979954 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979966 4972 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979979 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.979992 4972 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980006 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980019 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980048 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980060 4972 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980072 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980084 4972 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980096 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980109 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980123 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980137 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980149 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980162 4972 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980174 4972 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980186 4972 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980197 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980210 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980222 4972 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980235 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980248 4972 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980260 4972 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980272 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980285 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980298 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:12 crc kubenswrapper[4972]: I0228 10:37:12.980310 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.039150 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.039228 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.039268 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.039293 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.039310 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.070473 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.087554 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.096483 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 28 10:37:13 crc kubenswrapper[4972]: W0228 10:37:13.102163 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-d81565b42d19f83d52ee76597010664f213bfdde714083fc152635b67d33229e WatchSource:0}: Error finding container d81565b42d19f83d52ee76597010664f213bfdde714083fc152635b67d33229e: Status 404 returned error can't find the container with id d81565b42d19f83d52ee76597010664f213bfdde714083fc152635b67d33229e Feb 28 10:37:13 crc kubenswrapper[4972]: W0228 10:37:13.125330 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6922ecce55b52e8662443423336591e46ddb5ddc69c4afce366f1317680c3533 WatchSource:0}: Error finding container 6922ecce55b52e8662443423336591e46ddb5ddc69c4afce366f1317680c3533: Status 404 returned error can't find the container with id 6922ecce55b52e8662443423336591e46ddb5ddc69c4afce366f1317680c3533 Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.142655 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.142713 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.142728 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.142751 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.142766 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.225509 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6922ecce55b52e8662443423336591e46ddb5ddc69c4afce366f1317680c3533"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.227128 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d81565b42d19f83d52ee76597010664f213bfdde714083fc152635b67d33229e"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.228282 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9a7a46cf0446eb4e2d9d0c7b23dd2604cb489ab6083e72e50c8dc945d11a213a"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.246585 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.246632 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.246648 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.246672 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.246689 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.350055 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.350109 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.350120 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.350139 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.350151 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.386889 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.387063 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:14.387039964 +0000 UTC m=+91.299047702 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.454107 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.454163 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.454177 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.454195 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.454208 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.487890 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.487955 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.487982 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.488021 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488215 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488273 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488299 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488307 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488359 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488450 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488522 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488234 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488490 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:14.488368489 +0000 UTC m=+91.400376257 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488665 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:14.488625267 +0000 UTC m=+91.400633215 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488716 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:14.488695189 +0000 UTC m=+91.400703197 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.488758 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:14.48873853 +0000 UTC m=+91.400746508 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.557399 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.557506 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.557548 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.557576 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.557592 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.661447 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.661577 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.661620 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.661651 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.661669 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.764665 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.764759 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.764775 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.764802 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.764846 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.788821 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:13 crc kubenswrapper[4972]: E0228 10:37:13.789027 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.794247 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.795092 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.795818 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.796533 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.797141 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.797675 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.798302 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.798905 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.799616 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.800141 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.800762 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.801451 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.801997 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.803408 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.804898 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.806127 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.807734 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.808193 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.808602 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.809808 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.811400 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.812955 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.814526 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.815794 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.816958 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.817730 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.818751 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.819823 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.820604 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.821547 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.822273 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.823058 4972 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.823214 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.827725 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.828905 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.829977 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.831109 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.835326 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.836868 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.837683 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.838904 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.839599 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.840540 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.841145 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.842100 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.842734 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.843965 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.844603 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.845718 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.846521 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.847613 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.848178 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.848782 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.849978 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.850614 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.851663 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.852247 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.855030 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:13Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.867775 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.867851 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.867864 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.867883 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.867896 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.873477 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:13Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.891728 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:13Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.910314 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:13Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.970807 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.970876 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.970893 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.970919 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:13 crc kubenswrapper[4972]: I0228 10:37:13.970937 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:13Z","lastTransitionTime":"2026-02-28T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.074047 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.074107 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.074135 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.074165 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.074186 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.176841 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.176889 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.176900 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.176916 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.176929 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.234825 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4bb8e992fffd2c3e7d3ef3d3d6df5143baac8b007ae85acfddcaa47145736ae9"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.234906 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c22e413413dc7fb7423e209348674a7ee5dce77caccc46cf653bb5fe1abd8bfe"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.236663 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f517ff60ed44c14a79bb4454c5b8bf403b94df93a88eb814d50cf6e3540bd19c"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.271803 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67942065-2473-49bb-80b2-da52771cb5dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:36:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:36:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c6fec3e5d6ea3a2942df7fdc26a2abf4ab4931aef2ba4aa1cd818d603628a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2931cdd11c93734a665ce4fe41e0409614dd58eed794a630bc0169ccaddcfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://146b134a383e834cdaf63b60d1a64bb03e43c8f661b684ca208917ea82851b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c169fe4a9eb1ae7619edf132ee5f291127f15ee985f86587341337613dad49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b24aec46ddd08f8da7f9a2e56f33f704eeea057bda264637e05fe43d125e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-28T10:35:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.279197 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.279273 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.279292 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.279318 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.279336 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.292336 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.310020 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.323751 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bb8e992fffd2c3e7d3ef3d3d6df5143baac8b007ae85acfddcaa47145736ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c22e413413dc7fb7423e209348674a7ee5dce77caccc46cf653bb5fe1abd8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.338561 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.354574 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.370019 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.381803 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.381863 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.381876 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.381897 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.381913 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.390148 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.396731 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.396985 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:16.396961492 +0000 UTC m=+93.308969230 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.404374 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bb8e992fffd2c3e7d3ef3d3d6df5143baac8b007ae85acfddcaa47145736ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c22e413413dc7fb7423e209348674a7ee5dce77caccc46cf653bb5fe1abd8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.417316 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.430479 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.449652 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67942065-2473-49bb-80b2-da52771cb5dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:36:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:36:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-28T10:35:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c6fec3e5d6ea3a2942df7fdc26a2abf4ab4931aef2ba4aa1cd818d603628a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2931cdd11c93734a665ce4fe41e0409614dd58eed794a630bc0169ccaddcfbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://146b134a383e834cdaf63b60d1a64bb03e43c8f661b684ca208917ea82851b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c169fe4a9eb1ae7619edf132ee5f291127f15ee985f86587341337613dad49f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b24aec46ddd08f8da7f9a2e56f33f704eeea057bda264637e05fe43d125e84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:35:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b44f2c579244e740eb0ff26336602a7a6e7fcbf48ee1e00378d95e7c5f629250\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e8401dccc91edc68145968ea5f1ab046f51f85bf968a52fe76195d3eed5905a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81c2decb8c00349cb6e0831ef586c8fdf4bc7f62db5a8cd152edcd2eaa3f0089\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-28T10:35:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-28T10:35:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-28T10:35:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.462822 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f517ff60ed44c14a79bb4454c5b8bf403b94df93a88eb814d50cf6e3540bd19c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-28T10:37:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.480156 4972 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-28T10:37:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-28T10:37:14Z is after 2025-08-24T17:21:41Z" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.484248 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.484291 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.484302 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.484322 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.484334 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.497689 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.497741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.497761 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.497782 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497855 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497910 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:16.497895846 +0000 UTC m=+93.409903584 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497924 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497942 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498087 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:16.498044591 +0000 UTC m=+93.410052339 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497955 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498156 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.497924 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498218 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498232 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498289 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:16.498254768 +0000 UTC m=+93.410262536 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.498323 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:16.498309819 +0000 UTC m=+93.410317587 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.586572 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.586626 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.586636 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.586653 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.586663 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.689135 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.689206 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.689218 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.689242 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.689255 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.788594 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.788667 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.788746 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:14 crc kubenswrapper[4972]: E0228 10:37:14.788836 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.791026 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.791051 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.791061 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.791075 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.791086 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.894063 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.894118 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.894132 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.894155 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.894168 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.997151 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.997193 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.997209 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.997227 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:14 crc kubenswrapper[4972]: I0228 10:37:14.997239 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:14Z","lastTransitionTime":"2026-02-28T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.100870 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.100910 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.100921 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.100938 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.100951 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.204344 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.204402 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.204418 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.204445 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.204484 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.307238 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.307289 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.307303 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.307332 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.307353 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.410528 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.410572 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.410582 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.410600 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.410614 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.513206 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.513253 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.513267 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.513289 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.513302 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.615915 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.615972 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.615991 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.616017 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.616039 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.718227 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.718290 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.718307 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.718327 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.718340 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.788787 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:15 crc kubenswrapper[4972]: E0228 10:37:15.788978 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.820886 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.820942 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.820961 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.820997 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.821032 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.924837 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.924879 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.924889 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.924905 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:15 crc kubenswrapper[4972]: I0228 10:37:15.924916 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:15Z","lastTransitionTime":"2026-02-28T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.027536 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.027599 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.027616 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.027638 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.027651 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.132310 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.132358 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.132370 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.132389 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.132402 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.235305 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.235370 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.235385 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.235408 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.235425 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.243959 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9efe7d75b8e6a0dba623465e91e1c0388915c13e14f0ab648124ec61f25470c8"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.338577 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.338642 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.338652 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.338669 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.338680 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.393515 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.3934505169999998 podStartE2EDuration="3.393450517s" podCreationTimestamp="2026-02-28 10:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:16.393350904 +0000 UTC m=+93.305358682" watchObservedRunningTime="2026-02-28 10:37:16.393450517 +0000 UTC m=+93.305458265" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.416178 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.416531 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:20.416436429 +0000 UTC m=+97.328444207 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.441791 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.441844 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.441858 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.441874 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.441889 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.517767 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.517816 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.517843 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.517871 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.517993 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518031 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518033 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518060 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518087 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518096 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518118 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518062 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:20.518041414 +0000 UTC m=+97.430049162 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518155 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:20.518137477 +0000 UTC m=+97.430145225 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518179 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:20.518167197 +0000 UTC m=+97.430174945 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518211 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.518399 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:20.518359533 +0000 UTC m=+97.430367311 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.545351 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.545396 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.545405 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.545423 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.545436 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.648961 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.649038 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.649063 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.649090 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.649104 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.751299 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.751356 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.751369 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.751392 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.751405 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.788493 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.788712 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.788815 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:16 crc kubenswrapper[4972]: E0228 10:37:16.789020 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.854079 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.854127 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.854139 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.854167 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.854183 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.959359 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.959416 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.959429 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.959449 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:16 crc kubenswrapper[4972]: I0228 10:37:16.959479 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:16Z","lastTransitionTime":"2026-02-28T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.062857 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.062918 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.062932 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.062952 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.062964 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.166057 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.166122 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.166138 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.166160 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.166173 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.269118 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.269177 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.269190 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.269217 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.269232 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.371998 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.372052 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.372066 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.372088 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.372104 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.475597 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.475656 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.475668 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.475687 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.475703 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.580355 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.580490 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.580510 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.580541 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.580558 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.684494 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.684570 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.684595 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.684626 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.684646 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788447 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788556 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788592 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788641 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788685 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.788710 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: E0228 10:37:17.788770 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.810091 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.810114 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 28 10:37:17 crc kubenswrapper[4972]: E0228 10:37:17.810499 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.892255 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.892323 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.892336 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.892359 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.892373 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.995756 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.995871 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.995887 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.995920 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:17 crc kubenswrapper[4972]: I0228 10:37:17.995934 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:17Z","lastTransitionTime":"2026-02-28T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.098758 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.099073 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.099190 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.099296 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.099401 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.202295 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.202393 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.202425 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.202498 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.202521 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.249748 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:37:18 crc kubenswrapper[4972]: E0228 10:37:18.249991 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.305361 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.305442 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.305495 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.305530 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.305552 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.408905 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.408972 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.408997 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.409031 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.409072 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.513318 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.513386 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.513400 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.513425 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.513441 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.616938 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.617021 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.617042 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.617075 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.617297 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.720182 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.720253 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.720272 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.720300 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.720320 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.788247 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.788313 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:18 crc kubenswrapper[4972]: E0228 10:37:18.788453 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:18 crc kubenswrapper[4972]: E0228 10:37:18.788591 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.823729 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.823827 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.823852 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.823887 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.823915 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.928058 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.928133 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.928152 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.928184 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:18 crc kubenswrapper[4972]: I0228 10:37:18.928204 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:18Z","lastTransitionTime":"2026-02-28T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.030914 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.030974 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.030991 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.031013 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.031024 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.133373 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.133411 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.133421 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.133438 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.133451 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.236885 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.236929 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.236942 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.236959 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.236976 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.339986 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.340377 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.340492 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.340579 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.340652 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.443783 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.443841 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.443851 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.443871 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.443881 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.547154 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.547233 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.547247 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.547275 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.547292 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.649730 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.649805 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.649836 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.649857 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.649872 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.752847 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.752919 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.752932 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.752956 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.752968 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.788439 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:19 crc kubenswrapper[4972]: E0228 10:37:19.788661 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.856080 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.856152 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.856167 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.856195 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.856225 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.959565 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.959618 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.959629 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.959646 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:19 crc kubenswrapper[4972]: I0228 10:37:19.959662 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:19Z","lastTransitionTime":"2026-02-28T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.063214 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.063291 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.063311 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.063340 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.063361 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.166870 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.166929 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.166942 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.166963 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.166981 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.269854 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.269902 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.269913 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.269930 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.269942 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.372528 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.372576 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.372586 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.372601 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.372612 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.459724 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.460176 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:28.460120264 +0000 UTC m=+105.372128052 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.475568 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.475640 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.475658 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.475687 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.475704 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.560779 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.560834 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.560860 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.560885 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.560965 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.560967 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.560998 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561012 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561022 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:28.561005096 +0000 UTC m=+105.473012834 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561038 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561107 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561155 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561174 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561059 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:28.561035557 +0000 UTC m=+105.473043295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561257 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:28.561233803 +0000 UTC m=+105.473241731 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.561281 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:28.561268334 +0000 UTC m=+105.473276292 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.578894 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.578961 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.578980 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.579012 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.579031 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.681634 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.681683 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.681692 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.681707 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.681717 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.783600 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.783656 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.783673 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.783695 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.783714 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.788932 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.788976 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.789020 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:20 crc kubenswrapper[4972]: E0228 10:37:20.789147 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.886724 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.886809 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.886844 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.886875 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.886898 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.989769 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.989839 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.989880 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.989910 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:20 crc kubenswrapper[4972]: I0228 10:37:20.989931 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:20Z","lastTransitionTime":"2026-02-28T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.092877 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.092924 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.092937 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.092961 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.092973 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:21Z","lastTransitionTime":"2026-02-28T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.094147 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.094178 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.094189 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.094203 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.094214 4972 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-28T10:37:21Z","lastTransitionTime":"2026-02-28T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.789047 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:21 crc kubenswrapper[4972]: E0228 10:37:21.789198 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.876921 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 28 10:37:21 crc kubenswrapper[4972]: I0228 10:37:21.884733 4972 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 28 10:37:22 crc kubenswrapper[4972]: I0228 10:37:22.789033 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:22 crc kubenswrapper[4972]: I0228 10:37:22.789160 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:22 crc kubenswrapper[4972]: E0228 10:37:22.789219 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:22 crc kubenswrapper[4972]: E0228 10:37:22.789348 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:23 crc kubenswrapper[4972]: I0228 10:37:23.788843 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:23 crc kubenswrapper[4972]: E0228 10:37:23.789113 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:24 crc kubenswrapper[4972]: I0228 10:37:24.788115 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:24 crc kubenswrapper[4972]: I0228 10:37:24.788310 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:24 crc kubenswrapper[4972]: E0228 10:37:24.788429 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:24 crc kubenswrapper[4972]: E0228 10:37:24.788623 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:25 crc kubenswrapper[4972]: I0228 10:37:25.788901 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:25 crc kubenswrapper[4972]: E0228 10:37:25.789139 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:26 crc kubenswrapper[4972]: I0228 10:37:26.788590 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:26 crc kubenswrapper[4972]: I0228 10:37:26.788634 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:26 crc kubenswrapper[4972]: E0228 10:37:26.788808 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:26 crc kubenswrapper[4972]: E0228 10:37:26.788979 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:27 crc kubenswrapper[4972]: I0228 10:37:27.788751 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:27 crc kubenswrapper[4972]: E0228 10:37:27.788915 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.537073 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.537371 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.53731903 +0000 UTC m=+121.449326778 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.546249 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-pdjls"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.546891 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.548813 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.549536 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.549557 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.573715 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w2q4v"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.574560 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.574780 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zl2qc"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.576725 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.578501 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.580472 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.581298 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-pl4fx"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.581759 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584329 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584523 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584712 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584923 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584950 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.584998 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.585487 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.586893 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.588384 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.590192 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.623632 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-crbqv"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.624569 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.628234 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.628692 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.630321 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.631029 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.631288 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.631730 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.641235 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.642031 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.642087 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.642116 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.642151 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642256 4972 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642250 4972 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642322 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.642305034 +0000 UTC m=+121.554312772 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642352 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.642331186 +0000 UTC m=+121.554338924 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642468 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642483 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642495 4972 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642531 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.642523001 +0000 UTC m=+121.554530739 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642595 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642604 4972 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642612 4972 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.642638 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.642630665 +0000 UTC m=+121.554638403 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.728050 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.728633 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.731339 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.732011 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.732112 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.732136 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742650 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742689 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742707 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742739 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742760 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-multus\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742775 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-daemon-config\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742792 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-multus-certs\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742811 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742825 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742841 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/118d9c89-cc02-47b8-886e-e72848028ff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742859 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-cnibin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742875 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gztpv\" (UniqueName: \"kubernetes.io/projected/0302d7e4-875d-479b-9773-683e7f17a9f7-kube-api-access-gztpv\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742896 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742914 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-cni-binary-copy\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742931 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-conf-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742948 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.742967 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743033 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743058 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/118d9c89-cc02-47b8-886e-e72848028ff2-rootfs\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743085 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/118d9c89-cc02-47b8-886e-e72848028ff2-proxy-tls\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743103 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-os-release\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743124 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743140 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743160 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/43832a58-f393-4848-a2bb-fce9d2162d7a-hosts-file\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743188 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vw9k\" (UniqueName: \"kubernetes.io/projected/43832a58-f393-4848-a2bb-fce9d2162d7a-kube-api-access-5vw9k\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743283 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743326 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-netns\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743351 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743378 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743450 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-socket-dir-parent\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743492 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743510 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743530 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-system-cni-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743551 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtd9m\" (UniqueName: \"kubernetes.io/projected/9ef239ef-619d-4156-941b-4adc59bbdc2d-kube-api-access-xtd9m\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743573 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-k8s-cni-cncf-io\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743601 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-kubelet\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743622 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743639 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-hostroot\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743659 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-etc-kubernetes\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743681 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-system-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743732 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743761 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743800 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srhwm\" (UniqueName: \"kubernetes.io/projected/118d9c89-cc02-47b8-886e-e72848028ff2-kube-api-access-srhwm\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743832 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fknn\" (UniqueName: \"kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-cnibin\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743868 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-os-release\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743884 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743901 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743919 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.743949 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-bin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.788644 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.788700 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.788792 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:28 crc kubenswrapper[4972]: E0228 10:37:28.788901 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.822994 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8m6x5"] Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.823548 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.826729 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.826886 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.827239 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.827792 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844625 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-bin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844677 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800f3287-c3b6-43cc-88ec-e0e4737c2058-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844700 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844716 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844732 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844752 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-multus\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844785 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844780 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-bin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844822 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-daemon-config\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844844 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-multus-certs\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844855 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844885 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-cni-multus\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844865 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844912 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844855 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844960 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800f3287-c3b6-43cc-88ec-e0e4737c2058-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844973 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-multus-certs\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.844999 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/118d9c89-cc02-47b8-886e-e72848028ff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845121 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845166 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-cnibin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845208 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gztpv\" (UniqueName: \"kubernetes.io/projected/0302d7e4-875d-479b-9773-683e7f17a9f7-kube-api-access-gztpv\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845225 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845268 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845295 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-conf-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845324 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845348 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845447 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-cni-binary-copy\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845486 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845506 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/118d9c89-cc02-47b8-886e-e72848028ff2-rootfs\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845524 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/118d9c89-cc02-47b8-886e-e72848028ff2-proxy-tls\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845588 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-os-release\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845605 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845622 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845640 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vw9k\" (UniqueName: \"kubernetes.io/projected/43832a58-f393-4848-a2bb-fce9d2162d7a-kube-api-access-5vw9k\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845659 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845676 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-netns\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845697 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845716 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/43832a58-f393-4848-a2bb-fce9d2162d7a-hosts-file\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845742 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-daemon-config\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845788 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-socket-dir-parent\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845806 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845825 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845844 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845861 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/800f3287-c3b6-43cc-88ec-e0e4737c2058-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845874 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.845998 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/118d9c89-cc02-47b8-886e-e72848028ff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846110 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-os-release\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846129 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846140 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-netns\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846153 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-system-cni-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846149 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846177 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-system-cni-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846220 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-conf-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846217 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtd9m\" (UniqueName: \"kubernetes.io/projected/9ef239ef-619d-4156-941b-4adc59bbdc2d-kube-api-access-xtd9m\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846230 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846259 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846263 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-k8s-cni-cncf-io\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846235 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846158 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846290 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/43832a58-f393-4848-a2bb-fce9d2162d7a-hosts-file\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846240 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-multus-socket-dir-parent\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846291 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846321 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-run-k8s-cni-cncf-io\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846331 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846359 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-cnibin\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846367 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846383 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-kubelet\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846413 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/118d9c89-cc02-47b8-886e-e72848028ff2-rootfs\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846484 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-host-var-lib-kubelet\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846525 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-hostroot\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846602 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846659 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-etc-kubernetes\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846661 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-hostroot\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846691 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-system-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846687 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-etc-kubernetes\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846662 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846760 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846798 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-system-cni-dir\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846896 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846929 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846951 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srhwm\" (UniqueName: \"kubernetes.io/projected/118d9c89-cc02-47b8-886e-e72848028ff2-kube-api-access-srhwm\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846974 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-cnibin\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.846996 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-os-release\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847000 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847014 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847095 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0302d7e4-875d-479b-9773-683e7f17a9f7-os-release\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847105 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fknn\" (UniqueName: \"kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847049 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-cnibin\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847041 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847135 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847165 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847185 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847262 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847419 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9ef239ef-619d-4156-941b-4adc59bbdc2d-cni-binary-copy\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847633 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0302d7e4-875d-479b-9773-683e7f17a9f7-cni-binary-copy\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.847849 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9ef239ef-619d-4156-941b-4adc59bbdc2d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.852565 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/118d9c89-cc02-47b8-886e-e72848028ff2-proxy-tls\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.852931 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.865995 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gztpv\" (UniqueName: \"kubernetes.io/projected/0302d7e4-875d-479b-9773-683e7f17a9f7-kube-api-access-gztpv\") pod \"multus-pl4fx\" (UID: \"0302d7e4-875d-479b-9773-683e7f17a9f7\") " pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.867179 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fknn\" (UniqueName: \"kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn\") pod \"ovnkube-node-crbqv\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.869897 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vw9k\" (UniqueName: \"kubernetes.io/projected/43832a58-f393-4848-a2bb-fce9d2162d7a-kube-api-access-5vw9k\") pod \"node-resolver-pdjls\" (UID: \"43832a58-f393-4848-a2bb-fce9d2162d7a\") " pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.870117 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtd9m\" (UniqueName: \"kubernetes.io/projected/9ef239ef-619d-4156-941b-4adc59bbdc2d-kube-api-access-xtd9m\") pod \"multus-additional-cni-plugins-zl2qc\" (UID: \"9ef239ef-619d-4156-941b-4adc59bbdc2d\") " pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.874785 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srhwm\" (UniqueName: \"kubernetes.io/projected/118d9c89-cc02-47b8-886e-e72848028ff2-kube-api-access-srhwm\") pod \"machine-config-daemon-w2q4v\" (UID: \"118d9c89-cc02-47b8-886e-e72848028ff2\") " pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.889649 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.903581 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.908352 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pl4fx" Feb 28 10:37:28 crc kubenswrapper[4972]: W0228 10:37:28.915726 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ef239ef_619d_4156_941b_4adc59bbdc2d.slice/crio-0cb3b1d6f3531de732cbfd470fa83ffad6f20965f1bc8c437b3236dab38992b2 WatchSource:0}: Error finding container 0cb3b1d6f3531de732cbfd470fa83ffad6f20965f1bc8c437b3236dab38992b2: Status 404 returned error can't find the container with id 0cb3b1d6f3531de732cbfd470fa83ffad6f20965f1bc8c437b3236dab38992b2 Feb 28 10:37:28 crc kubenswrapper[4972]: W0228 10:37:28.921685 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0302d7e4_875d_479b_9773_683e7f17a9f7.slice/crio-39c49e03aa29864510393fd17f96cd365ee554b6eba9b05a8fd5d300b88f7003 WatchSource:0}: Error finding container 39c49e03aa29864510393fd17f96cd365ee554b6eba9b05a8fd5d300b88f7003: Status 404 returned error can't find the container with id 39c49e03aa29864510393fd17f96cd365ee554b6eba9b05a8fd5d300b88f7003 Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948448 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800f3287-c3b6-43cc-88ec-e0e4737c2058-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948511 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-host\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948603 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800f3287-c3b6-43cc-88ec-e0e4737c2058-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948770 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-586pm\" (UniqueName: \"kubernetes.io/projected/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-kube-api-access-586pm\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948824 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-serviceca\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/800f3287-c3b6-43cc-88ec-e0e4737c2058-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948893 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.949108 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.949301 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.948998 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/800f3287-c3b6-43cc-88ec-e0e4737c2058-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.949864 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/800f3287-c3b6-43cc-88ec-e0e4737c2058-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.950008 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.954210 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800f3287-c3b6-43cc-88ec-e0e4737c2058-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:28 crc kubenswrapper[4972]: I0228 10:37:28.969955 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800f3287-c3b6-43cc-88ec-e0e4737c2058-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vcjk2\" (UID: \"800f3287-c3b6-43cc-88ec-e0e4737c2058\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.005436 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29"] Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.006251 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.009385 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.009969 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.030208 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lcv58"] Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.030858 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.030954 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.047680 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.050290 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-586pm\" (UniqueName: \"kubernetes.io/projected/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-kube-api-access-586pm\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.050328 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-serviceca\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.050373 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-host\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.050440 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-host\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.053142 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-serviceca\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.073665 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-586pm\" (UniqueName: \"kubernetes.io/projected/b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0-kube-api-access-586pm\") pod \"node-ca-8m6x5\" (UID: \"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0\") " pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: W0228 10:37:29.080234 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod800f3287_c3b6_43cc_88ec_e0e4737c2058.slice/crio-de89683fd6adf3f5fe5102acd5ab1671622b052899fa5a51210a852a7c15856a WatchSource:0}: Error finding container de89683fd6adf3f5fe5102acd5ab1671622b052899fa5a51210a852a7c15856a: Status 404 returned error can't find the container with id de89683fd6adf3f5fe5102acd5ab1671622b052899fa5a51210a852a7c15856a Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.135639 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8m6x5" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151080 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48kms\" (UniqueName: \"kubernetes.io/projected/05bc9b1e-6461-4708-b2cb-261272230ffd-kube-api-access-48kms\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151199 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05bc9b1e-6461-4708-b2cb-261272230ffd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151224 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151246 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151265 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.151289 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfsxz\" (UniqueName: \"kubernetes.io/projected/b7f1c1c7-f8f6-418e-8526-e7946891f02e-kube-api-access-qfsxz\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.162560 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pdjls" Feb 28 10:37:29 crc kubenswrapper[4972]: W0228 10:37:29.174233 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5d0a4d1_f9ea_4044_8a33_0e3854e51fc0.slice/crio-c6e7078901377c71b245c31d485b3867c6285ffbb57644f50c7e617b47c6386c WatchSource:0}: Error finding container c6e7078901377c71b245c31d485b3867c6285ffbb57644f50c7e617b47c6386c: Status 404 returned error can't find the container with id c6e7078901377c71b245c31d485b3867c6285ffbb57644f50c7e617b47c6386c Feb 28 10:37:29 crc kubenswrapper[4972]: W0228 10:37:29.179527 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43832a58_f393_4848_a2bb_fce9d2162d7a.slice/crio-02e3c8d7534c9076c05659dec8807e73b666baeb5cdfbd7855fd13a177ed9d2b WatchSource:0}: Error finding container 02e3c8d7534c9076c05659dec8807e73b666baeb5cdfbd7855fd13a177ed9d2b: Status 404 returned error can't find the container with id 02e3c8d7534c9076c05659dec8807e73b666baeb5cdfbd7855fd13a177ed9d2b Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252239 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05bc9b1e-6461-4708-b2cb-261272230ffd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252319 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252357 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252410 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252445 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfsxz\" (UniqueName: \"kubernetes.io/projected/b7f1c1c7-f8f6-418e-8526-e7946891f02e-kube-api-access-qfsxz\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.252587 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48kms\" (UniqueName: \"kubernetes.io/projected/05bc9b1e-6461-4708-b2cb-261272230ffd-kube-api-access-48kms\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.253039 4972 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.253220 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs podName:b7f1c1c7-f8f6-418e-8526-e7946891f02e nodeName:}" failed. No retries permitted until 2026-02-28 10:37:29.753184856 +0000 UTC m=+106.665192634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs") pod "network-metrics-daemon-lcv58" (UID: "b7f1c1c7-f8f6-418e-8526-e7946891f02e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.254004 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.254145 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/05bc9b1e-6461-4708-b2cb-261272230ffd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.257909 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/05bc9b1e-6461-4708-b2cb-261272230ffd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.271413 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48kms\" (UniqueName: \"kubernetes.io/projected/05bc9b1e-6461-4708-b2cb-261272230ffd-kube-api-access-48kms\") pod \"ovnkube-control-plane-749d76644c-8tf29\" (UID: \"05bc9b1e-6461-4708-b2cb-261272230ffd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.275570 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfsxz\" (UniqueName: \"kubernetes.io/projected/b7f1c1c7-f8f6-418e-8526-e7946891f02e-kube-api-access-qfsxz\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.297823 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"cc2d8e083d238aa65b5c9cf64366286e34f6013e9da75d70d2a44cdbe5754e5d"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.297877 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.297903 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"2667dd69e1fcb4f5fc35f06465762fae2994b2c517f4c967601593d86eff5582"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.299189 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8m6x5" event={"ID":"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0","Type":"ContainerStarted","Data":"c6e7078901377c71b245c31d485b3867c6285ffbb57644f50c7e617b47c6386c"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.302138 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" exitCode=0 Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.302240 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.302304 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"e1546c6f07613d1f40ca78c524eedebc0e56d784610aeebb253389434f007566"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.306624 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pl4fx" event={"ID":"0302d7e4-875d-479b-9773-683e7f17a9f7","Type":"ContainerStarted","Data":"fdb2cc50d96375b8f53e4e976b0249d68f023875839c7bf59a00755f3c0209fe"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.306652 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pl4fx" event={"ID":"0302d7e4-875d-479b-9773-683e7f17a9f7","Type":"ContainerStarted","Data":"39c49e03aa29864510393fd17f96cd365ee554b6eba9b05a8fd5d300b88f7003"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.310032 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerStarted","Data":"f9038452c796a47d5a8d7e19a329d1723b7c943077d9d1a58879351d81c43168"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.310061 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerStarted","Data":"0cb3b1d6f3531de732cbfd470fa83ffad6f20965f1bc8c437b3236dab38992b2"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.314114 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podStartSLOduration=37.314061872 podStartE2EDuration="37.314061872s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:29.313566977 +0000 UTC m=+106.225574715" watchObservedRunningTime="2026-02-28 10:37:29.314061872 +0000 UTC m=+106.226069610" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.319134 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" event={"ID":"800f3287-c3b6-43cc-88ec-e0e4737c2058","Type":"ContainerStarted","Data":"db5f98929642b7d85a6b75053ceb8f5949fd6d9973fcc318d675db34e1bb3406"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.319233 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" event={"ID":"800f3287-c3b6-43cc-88ec-e0e4737c2058","Type":"ContainerStarted","Data":"de89683fd6adf3f5fe5102acd5ab1671622b052899fa5a51210a852a7c15856a"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.320909 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pdjls" event={"ID":"43832a58-f393-4848-a2bb-fce9d2162d7a","Type":"ContainerStarted","Data":"02e3c8d7534c9076c05659dec8807e73b666baeb5cdfbd7855fd13a177ed9d2b"} Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.329223 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.357283 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-pl4fx" podStartSLOduration=37.357258944 podStartE2EDuration="37.357258944s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:29.333189409 +0000 UTC m=+106.245197147" watchObservedRunningTime="2026-02-28 10:37:29.357258944 +0000 UTC m=+106.269266692" Feb 28 10:37:29 crc kubenswrapper[4972]: W0228 10:37:29.360181 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05bc9b1e_6461_4708_b2cb_261272230ffd.slice/crio-79e33edac7427b875d002797bad6cfcc97403e6dda6293a8e2e96fd15ae61181 WatchSource:0}: Error finding container 79e33edac7427b875d002797bad6cfcc97403e6dda6293a8e2e96fd15ae61181: Status 404 returned error can't find the container with id 79e33edac7427b875d002797bad6cfcc97403e6dda6293a8e2e96fd15ae61181 Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.398381 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vcjk2" podStartSLOduration=37.398363542 podStartE2EDuration="37.398363542s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:29.398025292 +0000 UTC m=+106.310033020" watchObservedRunningTime="2026-02-28 10:37:29.398363542 +0000 UTC m=+106.310371280" Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.758695 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.758912 4972 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.759007 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs podName:b7f1c1c7-f8f6-418e-8526-e7946891f02e nodeName:}" failed. No retries permitted until 2026-02-28 10:37:30.758959719 +0000 UTC m=+107.670967457 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs") pod "network-metrics-daemon-lcv58" (UID: "b7f1c1c7-f8f6-418e-8526-e7946891f02e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:29 crc kubenswrapper[4972]: I0228 10:37:29.788637 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:29 crc kubenswrapper[4972]: E0228 10:37:29.788824 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.329541 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" event={"ID":"05bc9b1e-6461-4708-b2cb-261272230ffd","Type":"ContainerStarted","Data":"7f79173104134d9a44808a80f2d187e4144f5623c83e304ae8354cb921f6de79"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.330095 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" event={"ID":"05bc9b1e-6461-4708-b2cb-261272230ffd","Type":"ContainerStarted","Data":"3a2faeeb8404014c9c4f5e25fcc524aee3410477778165366e3c6bb7abbed0dd"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.330117 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" event={"ID":"05bc9b1e-6461-4708-b2cb-261272230ffd","Type":"ContainerStarted","Data":"79e33edac7427b875d002797bad6cfcc97403e6dda6293a8e2e96fd15ae61181"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.332504 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pdjls" event={"ID":"43832a58-f393-4848-a2bb-fce9d2162d7a","Type":"ContainerStarted","Data":"16048e82f39b973a7eabaddb86c132592e4defbfabafcb1adeda09d83f77df4f"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.334789 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8m6x5" event={"ID":"b5d0a4d1-f9ea-4044-8a33-0e3854e51fc0","Type":"ContainerStarted","Data":"54c75923f23578f8cdc95e781981a249eb09cb200828a5286858fdd63ec67ac6"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.337895 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338741 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" exitCode=1 Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338814 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338867 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338880 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338891 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338902 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.338915 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.340714 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="f9038452c796a47d5a8d7e19a329d1723b7c943077d9d1a58879351d81c43168" exitCode=0 Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.341186 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"f9038452c796a47d5a8d7e19a329d1723b7c943077d9d1a58879351d81c43168"} Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.346999 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8tf29" podStartSLOduration=38.346979573 podStartE2EDuration="38.346979573s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:30.346031645 +0000 UTC m=+107.258039403" watchObservedRunningTime="2026-02-28 10:37:30.346979573 +0000 UTC m=+107.258987311" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.362790 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-pdjls" podStartSLOduration=39.362758321 podStartE2EDuration="39.362758321s" podCreationTimestamp="2026-02-28 10:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:30.362551095 +0000 UTC m=+107.274558843" watchObservedRunningTime="2026-02-28 10:37:30.362758321 +0000 UTC m=+107.274766059" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.417511 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8m6x5" podStartSLOduration=39.417491864 podStartE2EDuration="39.417491864s" podCreationTimestamp="2026-02-28 10:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:30.385234308 +0000 UTC m=+107.297242066" watchObservedRunningTime="2026-02-28 10:37:30.417491864 +0000 UTC m=+107.329499602" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.769733 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:30 crc kubenswrapper[4972]: E0228 10:37:30.770008 4972 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:30 crc kubenswrapper[4972]: E0228 10:37:30.770249 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs podName:b7f1c1c7-f8f6-418e-8526-e7946891f02e nodeName:}" failed. No retries permitted until 2026-02-28 10:37:32.770234238 +0000 UTC m=+109.682241966 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs") pod "network-metrics-daemon-lcv58" (UID: "b7f1c1c7-f8f6-418e-8526-e7946891f02e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.788527 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:30 crc kubenswrapper[4972]: E0228 10:37:30.788656 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.788658 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:30 crc kubenswrapper[4972]: I0228 10:37:30.788710 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:30 crc kubenswrapper[4972]: E0228 10:37:30.788885 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:30 crc kubenswrapper[4972]: E0228 10:37:30.789093 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:31 crc kubenswrapper[4972]: I0228 10:37:31.348166 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerStarted","Data":"ac82550a4a42e49e0c455a8e64ee316c156a31bd2497d35935673280cf947720"} Feb 28 10:37:31 crc kubenswrapper[4972]: I0228 10:37:31.789019 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:31 crc kubenswrapper[4972]: E0228 10:37:31.789161 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:31 crc kubenswrapper[4972]: I0228 10:37:31.790071 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.353886 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.354490 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.356094 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.357633 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54"} Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.358014 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.358756 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="ac82550a4a42e49e0c455a8e64ee316c156a31bd2497d35935673280cf947720" exitCode=0 Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.358784 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"ac82550a4a42e49e0c455a8e64ee316c156a31bd2497d35935673280cf947720"} Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.388954 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.388932257 podStartE2EDuration="15.388932257s" podCreationTimestamp="2026-02-28 10:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:32.386503545 +0000 UTC m=+109.298511303" watchObservedRunningTime="2026-02-28 10:37:32.388932257 +0000 UTC m=+109.300940005" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.788065 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.788075 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:32 crc kubenswrapper[4972]: E0228 10:37:32.788529 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:32 crc kubenswrapper[4972]: E0228 10:37:32.788546 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.788065 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:32 crc kubenswrapper[4972]: E0228 10:37:32.788648 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:32 crc kubenswrapper[4972]: I0228 10:37:32.796062 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:32 crc kubenswrapper[4972]: E0228 10:37:32.796192 4972 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:32 crc kubenswrapper[4972]: E0228 10:37:32.796246 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs podName:b7f1c1c7-f8f6-418e-8526-e7946891f02e nodeName:}" failed. No retries permitted until 2026-02-28 10:37:36.796229789 +0000 UTC m=+113.708237527 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs") pod "network-metrics-daemon-lcv58" (UID: "b7f1c1c7-f8f6-418e-8526-e7946891f02e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:33 crc kubenswrapper[4972]: I0228 10:37:33.369205 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="dfbe1fa26cd43f06a840549eb733a9eb0e44b29be25c43e0db6251f91ffb0d3f" exitCode=0 Feb 28 10:37:33 crc kubenswrapper[4972]: I0228 10:37:33.369305 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"dfbe1fa26cd43f06a840549eb733a9eb0e44b29be25c43e0db6251f91ffb0d3f"} Feb 28 10:37:33 crc kubenswrapper[4972]: I0228 10:37:33.788888 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:33 crc kubenswrapper[4972]: E0228 10:37:33.789640 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:34 crc kubenswrapper[4972]: I0228 10:37:34.375388 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="d3bf04483daca2d9a22d23c937ba8c4e7b3c812d66a47ef3201caf31975e8408" exitCode=0 Feb 28 10:37:34 crc kubenswrapper[4972]: I0228 10:37:34.375440 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"d3bf04483daca2d9a22d23c937ba8c4e7b3c812d66a47ef3201caf31975e8408"} Feb 28 10:37:34 crc kubenswrapper[4972]: I0228 10:37:34.788612 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:34 crc kubenswrapper[4972]: I0228 10:37:34.788612 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:34 crc kubenswrapper[4972]: E0228 10:37:34.789016 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:34 crc kubenswrapper[4972]: E0228 10:37:34.789140 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:34 crc kubenswrapper[4972]: I0228 10:37:34.788601 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:34 crc kubenswrapper[4972]: E0228 10:37:34.789210 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.386609 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.388211 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.388607 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.388992 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.393645 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerStarted","Data":"3a43b883783b63c8a4ac7903521839ecadcab4ef97d81f4de15555c5fa7e4475"} Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.516435 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:35 crc kubenswrapper[4972]: I0228 10:37:35.789698 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:35 crc kubenswrapper[4972]: E0228 10:37:35.790358 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.401822 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.402756 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerStarted","Data":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.402864 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.402892 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.407968 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="3a43b883783b63c8a4ac7903521839ecadcab4ef97d81f4de15555c5fa7e4475" exitCode=0 Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.408006 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"3a43b883783b63c8a4ac7903521839ecadcab4ef97d81f4de15555c5fa7e4475"} Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.433810 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.455148 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podStartSLOduration=44.455113978 podStartE2EDuration="44.455113978s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:36.434846817 +0000 UTC m=+113.346854595" watchObservedRunningTime="2026-02-28 10:37:36.455113978 +0000 UTC m=+113.367121756" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.788402 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.788407 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.788575 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:36 crc kubenswrapper[4972]: E0228 10:37:36.788578 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:36 crc kubenswrapper[4972]: E0228 10:37:36.788830 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:36 crc kubenswrapper[4972]: E0228 10:37:36.788998 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:36 crc kubenswrapper[4972]: I0228 10:37:36.836920 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:36 crc kubenswrapper[4972]: E0228 10:37:36.837095 4972 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:36 crc kubenswrapper[4972]: E0228 10:37:36.837156 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs podName:b7f1c1c7-f8f6-418e-8526-e7946891f02e nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.837138291 +0000 UTC m=+121.749146029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs") pod "network-metrics-daemon-lcv58" (UID: "b7f1c1c7-f8f6-418e-8526-e7946891f02e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 28 10:37:37 crc kubenswrapper[4972]: I0228 10:37:37.261405 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lcv58"] Feb 28 10:37:37 crc kubenswrapper[4972]: I0228 10:37:37.415543 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ef239ef-619d-4156-941b-4adc59bbdc2d" containerID="7082c7a079c31463dd983439f8139353bbd7cbf78bcc81d358c39602c9245b0c" exitCode=0 Feb 28 10:37:37 crc kubenswrapper[4972]: I0228 10:37:37.415608 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerDied","Data":"7082c7a079c31463dd983439f8139353bbd7cbf78bcc81d358c39602c9245b0c"} Feb 28 10:37:37 crc kubenswrapper[4972]: I0228 10:37:37.415958 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:37 crc kubenswrapper[4972]: E0228 10:37:37.416705 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:37 crc kubenswrapper[4972]: I0228 10:37:37.789264 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:37 crc kubenswrapper[4972]: E0228 10:37:37.789503 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:38 crc kubenswrapper[4972]: I0228 10:37:38.421992 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" event={"ID":"9ef239ef-619d-4156-941b-4adc59bbdc2d","Type":"ContainerStarted","Data":"f874ff6feacc07a2c100b28922c0aa8f99fad6c7fabdc34564770c4562559d21"} Feb 28 10:37:38 crc kubenswrapper[4972]: I0228 10:37:38.446017 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zl2qc" podStartSLOduration=46.445993346 podStartE2EDuration="46.445993346s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:38.444873313 +0000 UTC m=+115.356881121" watchObservedRunningTime="2026-02-28 10:37:38.445993346 +0000 UTC m=+115.358001114" Feb 28 10:37:38 crc kubenswrapper[4972]: I0228 10:37:38.788528 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:38 crc kubenswrapper[4972]: I0228 10:37:38.788595 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:38 crc kubenswrapper[4972]: I0228 10:37:38.788551 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:38 crc kubenswrapper[4972]: E0228 10:37:38.788714 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 28 10:37:38 crc kubenswrapper[4972]: E0228 10:37:38.788841 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 28 10:37:38 crc kubenswrapper[4972]: E0228 10:37:38.788990 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lcv58" podUID="b7f1c1c7-f8f6-418e-8526-e7946891f02e" Feb 28 10:37:39 crc kubenswrapper[4972]: I0228 10:37:39.789117 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:39 crc kubenswrapper[4972]: E0228 10:37:39.789312 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 28 10:37:39 crc kubenswrapper[4972]: I0228 10:37:39.965060 4972 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 28 10:37:39 crc kubenswrapper[4972]: I0228 10:37:39.965212 4972 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.009012 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.009661 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.009723 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.026123 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.027523 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.032280 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.032583 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.032668 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.033162 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.033274 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.033311 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.039611 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.040184 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.041040 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5d29s"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.041470 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.041670 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpbj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.042031 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.043186 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.043558 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.043773 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mfrtc"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.044211 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.042455 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.046663 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7gn5m"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.047188 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.047602 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xwmd"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.048217 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.048703 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.048990 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052133 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052199 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052266 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052292 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052651 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052884 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.052995 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.053104 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.053755 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.053806 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.053928 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054040 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054053 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054142 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054181 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054211 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054274 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054305 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054348 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054365 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054218 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.054435 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.056598 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jp6lb"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.057122 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.057310 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.057381 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bd989"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.057797 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9f545"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.058072 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.058416 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.058454 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.058832 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.059248 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.059548 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.059557 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.060993 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.061588 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.061865 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.065798 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.065949 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.066042 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.077950 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.078824 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.079336 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.079404 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.079564 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.079717 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.079820 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.082290 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.082330 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.082354 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7dr\" (UniqueName: \"kubernetes.io/projected/398facc3-3546-42c1-8aac-07a59c92852c-kube-api-access-tf7dr\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.082443 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-images\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.087775 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.087918 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18b36f3e-aaad-4381-bf0f-af34a2c10337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.087969 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089512 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6bld\" (UniqueName: \"kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089557 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0f314b2-1a1a-4235-841d-28872c4907ba-serving-cert\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089590 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kjqt\" (UniqueName: \"kubernetes.io/projected/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-kube-api-access-7kjqt\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089609 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089641 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089710 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089828 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089890 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089919 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-config\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089957 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/964ea0e3-fd84-4e16-873b-b3013209243d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.089998 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090041 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090066 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090123 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090148 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090168 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-encryption-config\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090218 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090236 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb2eeb1-2033-4895-af78-9cf70134182e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090291 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk2gq\" (UniqueName: \"kubernetes.io/projected/3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53-kube-api-access-nk2gq\") pod \"downloads-7954f5f757-mfrtc\" (UID: \"3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53\") " pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090319 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrv79\" (UniqueName: \"kubernetes.io/projected/8b62a70e-5da5-48a8-9eae-e5b529606f3c-kube-api-access-qrv79\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090361 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090382 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090400 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-policies\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090435 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b62a70e-5da5-48a8-9eae-e5b529606f3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090456 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090510 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgtr9\" (UniqueName: \"kubernetes.io/projected/9bb2eeb1-2033-4895-af78-9cf70134182e-kube-api-access-kgtr9\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090534 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-service-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.090570 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgllx\" (UniqueName: \"kubernetes.io/projected/c6085b35-69c1-40e1-a782-d9f7f5708181-kube-api-access-xgllx\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091529 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9q7w\" (UniqueName: \"kubernetes.io/projected/5d5efe87-5fb1-442e-b7bd-5c9758daa505-kube-api-access-t9q7w\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091580 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/398facc3-3546-42c1-8aac-07a59c92852c-serving-cert\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091603 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjt7\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-kube-api-access-lmjt7\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091623 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d5efe87-5fb1-442e-b7bd-5c9758daa505-machine-approver-tls\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091644 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xshqs\" (UniqueName: \"kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091708 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091732 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091752 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-config\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091772 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091796 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ktd2\" (UniqueName: \"kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091814 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxwcp\" (UniqueName: \"kubernetes.io/projected/18b36f3e-aaad-4381-bf0f-af34a2c10337-kube-api-access-nxwcp\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.091974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-auth-proxy-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092005 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnpft\" (UniqueName: \"kubernetes.io/projected/3c1531f8-6f83-4253-a88f-b2ec8144f50d-kube-api-access-gnpft\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092026 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb2eeb1-2033-4895-af78-9cf70134182e-serving-cert\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092053 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092074 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092119 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092139 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b62a70e-5da5-48a8-9eae-e5b529606f3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092168 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-client\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092192 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-dir\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092213 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-trusted-ca\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092698 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.092893 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.093022 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.093026 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.093604 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.093667 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.093929 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.094480 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.104248 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2xkbh"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105571 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105709 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105849 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105905 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-config\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105946 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb996\" (UniqueName: \"kubernetes.io/projected/b0f314b2-1a1a-4235-841d-28872c4907ba-kube-api-access-vb996\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105973 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd06083-d0dc-4523-b832-471ca8b45e82-metrics-tls\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.105993 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w8sp\" (UniqueName: \"kubernetes.io/projected/fcd06083-d0dc-4523-b832-471ca8b45e82-kube-api-access-8w8sp\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106015 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-client\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106043 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/964ea0e3-fd84-4e16-873b-b3013209243d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106062 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-serving-cert\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106085 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106100 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j82kg\" (UniqueName: \"kubernetes.io/projected/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-kube-api-access-j82kg\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106118 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18b36f3e-aaad-4381-bf0f-af34a2c10337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106135 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106154 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106171 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106191 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106207 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-config\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106223 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-serving-cert\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106237 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.106253 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6085b35-69c1-40e1-a782-d9f7f5708181-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.107214 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.107404 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.107922 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.108184 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.109159 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.109250 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.109361 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.109542 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.109763 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.110147 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.110337 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.110427 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.110332 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.110352 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.108543 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.111178 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.114012 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.114699 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.114769 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.114920 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115378 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115523 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115677 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115755 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115848 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115928 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116029 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116115 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116221 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116386 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116593 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116761 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.116904 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.117042 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115854 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115874 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.115897 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.117525 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.118633 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.118888 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.120194 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.121419 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.121819 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.123309 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.124042 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.127047 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-599cv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.128354 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.128429 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.135297 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.135795 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xstmm"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.136698 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.138922 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.145294 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.148822 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.151888 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-4ftt6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.152232 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.152557 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.158986 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.159256 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.159437 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.159553 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.159759 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.159818 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.161477 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.188771 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.189126 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.189816 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.190007 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.190500 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.190583 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.190791 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.192794 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-44c45"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.193926 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.194600 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.194971 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.195655 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.195702 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.202868 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.204242 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.204843 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.205201 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.205305 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.205395 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.206149 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.206486 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.206811 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207199 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-encryption-config\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207249 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207273 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnpd\" (UniqueName: \"kubernetes.io/projected/0389a687-8bad-46df-ac25-895f961d0184-kube-api-access-grnpd\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207298 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207321 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207358 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb2eeb1-2033-4895-af78-9cf70134182e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207388 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207611 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrv79\" (UniqueName: \"kubernetes.io/projected/8b62a70e-5da5-48a8-9eae-e5b529606f3c-kube-api-access-qrv79\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.207667 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk2gq\" (UniqueName: \"kubernetes.io/projected/3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53-kube-api-access-nk2gq\") pod \"downloads-7954f5f757-mfrtc\" (UID: \"3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53\") " pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.208238 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.208197 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.208532 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209130 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209282 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209332 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209353 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-policies\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209415 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209448 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-image-import-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209485 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb2eeb1-2033-4895-af78-9cf70134182e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209529 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b62a70e-5da5-48a8-9eae-e5b529606f3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209562 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209591 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-client\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.209599 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.210040 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.210642 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.210775 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-policies\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211106 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.210643 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgtr9\" (UniqueName: \"kubernetes.io/projected/9bb2eeb1-2033-4895-af78-9cf70134182e-kube-api-access-kgtr9\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211418 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-encryption-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211492 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-service-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211544 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgllx\" (UniqueName: \"kubernetes.io/projected/c6085b35-69c1-40e1-a782-d9f7f5708181-kube-api-access-xgllx\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211573 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgzgn\" (UniqueName: \"kubernetes.io/projected/8b782f84-114b-40d3-b7ba-806e1fccc04f-kube-api-access-xgzgn\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211634 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9q7w\" (UniqueName: \"kubernetes.io/projected/5d5efe87-5fb1-442e-b7bd-5c9758daa505-kube-api-access-t9q7w\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211680 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211693 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211734 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/398facc3-3546-42c1-8aac-07a59c92852c-serving-cert\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.211767 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjt7\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-kube-api-access-lmjt7\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212338 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-encryption-config\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212381 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212620 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-serving-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212652 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b637c49c-9ebf-40bc-8208-b3da5a05b4be-config\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212677 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d5efe87-5fb1-442e-b7bd-5c9758daa505-machine-approver-tls\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212690 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-service-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212693 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xshqs\" (UniqueName: \"kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212739 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212763 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212785 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-config\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212808 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212836 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212861 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0389a687-8bad-46df-ac25-895f961d0184-signing-cabundle\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212884 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxwcp\" (UniqueName: \"kubernetes.io/projected/18b36f3e-aaad-4381-bf0f-af34a2c10337-kube-api-access-nxwcp\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212908 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ktd2\" (UniqueName: \"kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212930 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-node-pullsecrets\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212946 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212955 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.212981 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b637c49c-9ebf-40bc-8208-b3da5a05b4be-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213002 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213031 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-metrics-certs\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213057 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-auth-proxy-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213082 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnpft\" (UniqueName: \"kubernetes.io/projected/3c1531f8-6f83-4253-a88f-b2ec8144f50d-kube-api-access-gnpft\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213105 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb2eeb1-2033-4895-af78-9cf70134182e-serving-cert\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213130 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b782f84-114b-40d3-b7ba-806e1fccc04f-service-ca-bundle\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213162 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213191 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213201 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.213920 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-config\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214014 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a03f59e-a71b-49d2-9a93-59075788fc21-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214070 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b637c49c-9ebf-40bc-8208-b3da5a05b4be-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214113 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214164 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdml8\" (UniqueName: \"kubernetes.io/projected/cc45436c-617d-4941-a3ab-952c1bfd4d4f-kube-api-access-pdml8\") pod \"migrator-59844c95c7-bs4ks\" (UID: \"cc45436c-617d-4941-a3ab-952c1bfd4d4f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214191 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-default-certificate\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.214999 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-auth-proxy-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.215504 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.215697 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.215920 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.215963 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/398facc3-3546-42c1-8aac-07a59c92852c-serving-cert\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216077 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mfrtc"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216074 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b62a70e-5da5-48a8-9eae-e5b529606f3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216112 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-client\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216128 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-dir\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216184 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216437 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-trusted-ca\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216479 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216503 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-audit-dir\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216531 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216762 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216783 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5d5efe87-5fb1-442e-b7bd-5c9758daa505-machine-approver-tls\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216821 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-config\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216890 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-serving-cert\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216919 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-client\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216968 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/964ea0e3-fd84-4e16-873b-b3013209243d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.216995 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bj6\" (UniqueName: \"kubernetes.io/projected/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-kube-api-access-68bj6\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217020 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-serving-cert\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217046 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb996\" (UniqueName: \"kubernetes.io/projected/b0f314b2-1a1a-4235-841d-28872c4907ba-kube-api-access-vb996\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd06083-d0dc-4523-b832-471ca8b45e82-metrics-tls\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217157 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w8sp\" (UniqueName: \"kubernetes.io/projected/fcd06083-d0dc-4523-b832-471ca8b45e82-kube-api-access-8w8sp\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217182 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217205 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j82kg\" (UniqueName: \"kubernetes.io/projected/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-kube-api-access-j82kg\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217228 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18b36f3e-aaad-4381-bf0f-af34a2c10337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217244 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217578 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217596 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.217744 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/398facc3-3546-42c1-8aac-07a59c92852c-trusted-ca\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.218834 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b62a70e-5da5-48a8-9eae-e5b529606f3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.218958 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/964ea0e3-fd84-4e16-873b-b3013209243d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.219522 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-config\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.219563 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5d29s"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.219843 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220215 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd06083-d0dc-4523-b832-471ca8b45e82-metrics-tls\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220492 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220553 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220573 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220609 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.220625 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-config\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.221528 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-etcd-client\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.221735 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-client\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.221772 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222094 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222531 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222663 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222770 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6085b35-69c1-40e1-a782-d9f7f5708181-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222870 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-serving-cert\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.222980 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f314b2-1a1a-4235-841d-28872c4907ba-config\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.223033 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18b36f3e-aaad-4381-bf0f-af34a2c10337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.223286 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.223530 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0389a687-8bad-46df-ac25-895f961d0184-signing-key\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.224559 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1531f8-6f83-4253-a88f-b2ec8144f50d-serving-cert\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.224688 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.224756 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225193 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3c1531f8-6f83-4253-a88f-b2ec8144f50d-etcd-ca\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225334 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a03f59e-a71b-49d2-9a93-59075788fc21-config\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225376 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-images\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225405 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225431 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-serving-cert\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225519 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225546 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7dr\" (UniqueName: \"kubernetes.io/projected/398facc3-3546-42c1-8aac-07a59c92852c-kube-api-access-tf7dr\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225576 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18b36f3e-aaad-4381-bf0f-af34a2c10337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225601 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225628 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl659\" (UniqueName: \"kubernetes.io/projected/40d5e1c5-21a4-419a-9425-6773e3f33c29-kube-api-access-gl659\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225649 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kjqt\" (UniqueName: \"kubernetes.io/projected/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-kube-api-access-7kjqt\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225673 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.225709 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6bld\" (UniqueName: \"kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226371 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-images\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226521 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0f314b2-1a1a-4235-841d-28872c4907ba-serving-cert\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226550 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226574 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226596 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69btw\" (UniqueName: \"kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226617 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.226623 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227272 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227386 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227411 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227432 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-config\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227476 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227490 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227511 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-config\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227528 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/964ea0e3-fd84-4e16-873b-b3013209243d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227546 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227565 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a03f59e-a71b-49d2-9a93-59075788fc21-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227601 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-stats-auth\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.227836 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-serving-cert\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.228409 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18b36f3e-aaad-4381-bf0f-af34a2c10337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.229025 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.229268 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.229303 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.229337 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit-dir\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.229929 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.231067 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5efe87-5fb1-442e-b7bd-5c9758daa505-config\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.231595 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb2eeb1-2033-4895-af78-9cf70134182e-serving-cert\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.231974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.232416 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.232742 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.232887 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpbj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.233395 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xwmd"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.234268 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-config\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.234956 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7gn5m"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.235329 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6085b35-69c1-40e1-a782-d9f7f5708181-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.235445 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.240554 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/964ea0e3-fd84-4e16-873b-b3013209243d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.240728 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.242594 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.254396 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.255305 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.256613 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xstmm"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.258398 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.262362 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0f314b2-1a1a-4235-841d-28872c4907ba-serving-cert\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.262649 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.267522 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.269005 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.273317 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.276504 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jnb9w"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.279061 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.279218 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.282495 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2xkbh"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.283940 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9f545"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.287638 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.291645 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.291693 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-44c45"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.292222 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.292335 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.292595 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b62a70e-5da5-48a8-9eae-e5b529606f3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.293098 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.293313 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.294451 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bd989"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.300233 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.301293 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.302203 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.303150 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jp6lb"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.304232 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-599cv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.305191 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-5plg9"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.306057 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.306172 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.307427 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.307586 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nkn8q"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.309073 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.310407 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.310671 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.311544 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.312446 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.313527 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.314438 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.315622 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.316406 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jnb9w"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.317681 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nkn8q"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.318609 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.319556 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zxcst"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.320752 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-nfhbz"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.321323 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.321888 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zxcst"] Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.322595 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.327677 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.330402 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wddsg\" (UniqueName: \"kubernetes.io/projected/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-kube-api-access-wddsg\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.330642 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0389a687-8bad-46df-ac25-895f961d0184-signing-key\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.330817 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63b91360-360a-46bf-ab58-6124f4210b84-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.331000 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl659\" (UniqueName: \"kubernetes.io/projected/40d5e1c5-21a4-419a-9425-6773e3f33c29-kube-api-access-gl659\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.331179 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.332070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69btw\" (UniqueName: \"kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.332249 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.332410 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-stats-auth\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.332650 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.332841 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63b91360-360a-46bf-ab58-6124f4210b84-proxy-tls\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333094 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7k4\" (UniqueName: \"kubernetes.io/projected/79293b7f-3383-4376-90ca-1f6c6eff2dc4-kube-api-access-5f7k4\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333296 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-client\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333335 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-encryption-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333365 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333418 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b637c49c-9ebf-40bc-8208-b3da5a05b4be-config\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333441 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333482 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zwgg\" (UniqueName: \"kubernetes.io/projected/043d7cc4-5732-49b8-8150-bb897c5fa424-kube-api-access-7zwgg\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333514 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-node-pullsecrets\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333533 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333571 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b637c49c-9ebf-40bc-8208-b3da5a05b4be-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333590 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-metrics-certs\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333619 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a03f59e-a71b-49d2-9a93-59075788fc21-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333639 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdml8\" (UniqueName: \"kubernetes.io/projected/cc45436c-617d-4941-a3ab-952c1bfd4d4f-kube-api-access-pdml8\") pod \"migrator-59844c95c7-bs4ks\" (UID: \"cc45436c-617d-4941-a3ab-952c1bfd4d4f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333661 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-proxy-tls\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333684 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-serving-cert\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333712 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bj6\" (UniqueName: \"kubernetes.io/projected/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-kube-api-access-68bj6\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333742 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmhbq\" (UniqueName: \"kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333768 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a03f59e-a71b-49d2-9a93-59075788fc21-config\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333787 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333813 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-serving-cert\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333850 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333872 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-config\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333891 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a03f59e-a71b-49d2-9a93-59075788fc21-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333911 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit-dir\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333928 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnpd\" (UniqueName: \"kubernetes.io/projected/0389a687-8bad-46df-ac25-895f961d0184-kube-api-access-grnpd\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333949 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-images\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333977 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.333994 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/043d7cc4-5732-49b8-8150-bb897c5fa424-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334013 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-image-import-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334029 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-srv-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334059 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgzgn\" (UniqueName: \"kubernetes.io/projected/8b782f84-114b-40d3-b7ba-806e1fccc04f-kube-api-access-xgzgn\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334077 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334101 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-serving-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334128 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0389a687-8bad-46df-ac25-895f961d0184-signing-cabundle\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334156 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zchf\" (UniqueName: \"kubernetes.io/projected/63b91360-360a-46bf-ab58-6124f4210b84-kube-api-access-4zchf\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334187 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b782f84-114b-40d3-b7ba-806e1fccc04f-service-ca-bundle\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334210 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b637c49c-9ebf-40bc-8208-b3da5a05b4be-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334227 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnjz7\" (UniqueName: \"kubernetes.io/projected/06ae4169-c713-4ebc-88aa-dee750386952-kube-api-access-gnjz7\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334223 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334245 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-default-certificate\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334370 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334389 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit-dir\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.334517 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/40d5e1c5-21a4-419a-9425-6773e3f33c29-node-pullsecrets\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.335924 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-serving-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.335936 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.335989 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-audit\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.336222 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/40d5e1c5-21a4-419a-9425-6773e3f33c29-image-import-ca\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.337650 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-etcd-client\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.338868 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-encryption-config\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.338960 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.339575 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40d5e1c5-21a4-419a-9425-6773e3f33c29-serving-cert\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.348573 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.353760 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.368641 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.375851 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.388511 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.428171 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436418 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-images\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436538 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/043d7cc4-5732-49b8-8150-bb897c5fa424-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436584 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-srv-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436701 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436793 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zchf\" (UniqueName: \"kubernetes.io/projected/63b91360-360a-46bf-ab58-6124f4210b84-kube-api-access-4zchf\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436875 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnjz7\" (UniqueName: \"kubernetes.io/projected/06ae4169-c713-4ebc-88aa-dee750386952-kube-api-access-gnjz7\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.436953 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437001 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wddsg\" (UniqueName: \"kubernetes.io/projected/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-kube-api-access-wddsg\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437050 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63b91360-360a-46bf-ab58-6124f4210b84-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437102 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437183 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63b91360-360a-46bf-ab58-6124f4210b84-proxy-tls\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437221 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7k4\" (UniqueName: \"kubernetes.io/projected/79293b7f-3383-4376-90ca-1f6c6eff2dc4-kube-api-access-5f7k4\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437257 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437339 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zwgg\" (UniqueName: \"kubernetes.io/projected/043d7cc4-5732-49b8-8150-bb897c5fa424-kube-api-access-7zwgg\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437517 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-proxy-tls\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437658 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.437706 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmhbq\" (UniqueName: \"kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.438528 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.438579 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63b91360-360a-46bf-ab58-6124f4210b84-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.447983 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.468357 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.479092 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a03f59e-a71b-49d2-9a93-59075788fc21-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.488583 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.494850 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a03f59e-a71b-49d2-9a93-59075788fc21-config\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.507879 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.527771 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.554573 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.555991 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0389a687-8bad-46df-ac25-895f961d0184-signing-cabundle\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.568687 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.574649 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0389a687-8bad-46df-ac25-895f961d0184-signing-key\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.590081 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.608871 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.628896 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.638973 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b637c49c-9ebf-40bc-8208-b3da5a05b4be-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.647812 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.668718 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.676300 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b637c49c-9ebf-40bc-8208-b3da5a05b4be-config\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.688935 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.708172 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.719114 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-serving-cert\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.728587 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.747898 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.755047 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-config\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.769940 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.787096 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.788342 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.788519 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.788645 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.796300 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-stats-auth\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.808051 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.827399 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.840219 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-metrics-certs\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.847858 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.868022 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.877149 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b782f84-114b-40d3-b7ba-806e1fccc04f-service-ca-bundle\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.888164 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.908177 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.918099 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b782f84-114b-40d3-b7ba-806e1fccc04f-default-certificate\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.929022 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.947517 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.967607 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 28 10:37:40 crc kubenswrapper[4972]: I0228 10:37:40.987525 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.007330 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.028269 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.047567 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.068316 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.088245 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.108572 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.128251 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.148356 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.160956 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.168186 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.180737 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06ae4169-c713-4ebc-88aa-dee750386952-srv-cert\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.188731 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.207152 4972 request.go:700] Waited for 1.015885025s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.208870 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.236259 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.248627 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.261012 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/043d7cc4-5732-49b8-8150-bb897c5fa424-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.269180 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.277866 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-images\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.288815 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.308204 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.322018 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-proxy-tls\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.327697 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.348567 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.363072 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63b91360-360a-46bf-ab58-6124f4210b84-proxy-tls\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.369178 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.388075 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.402754 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.417756 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.428166 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.428583 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: E0228 10:37:41.437503 4972 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 28 10:37:41 crc kubenswrapper[4972]: E0228 10:37:41.437637 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert podName:79293b7f-3383-4376-90ca-1f6c6eff2dc4 nodeName:}" failed. No retries permitted until 2026-02-28 10:37:41.937611751 +0000 UTC m=+118.849619709 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-2fwv6" (UID: "79293b7f-3383-4376-90ca-1f6c6eff2dc4") : failed to sync secret cache: timed out waiting for the condition Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.448504 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.468516 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.488921 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.508152 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.535638 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.548259 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.567625 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.588113 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.607931 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.667846 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk2gq\" (UniqueName: \"kubernetes.io/projected/3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53-kube-api-access-nk2gq\") pod \"downloads-7954f5f757-mfrtc\" (UID: \"3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53\") " pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.668962 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.707846 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.708657 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgtr9\" (UniqueName: \"kubernetes.io/projected/9bb2eeb1-2033-4895-af78-9cf70134182e-kube-api-access-kgtr9\") pod \"openshift-config-operator-7777fb866f-bd989\" (UID: \"9bb2eeb1-2033-4895-af78-9cf70134182e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.728501 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.749112 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.772649 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrv79\" (UniqueName: \"kubernetes.io/projected/8b62a70e-5da5-48a8-9eae-e5b529606f3c-kube-api-access-qrv79\") pod \"openshift-apiserver-operator-796bbdcf4f-lmtss\" (UID: \"8b62a70e-5da5-48a8-9eae-e5b529606f3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.786002 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgllx\" (UniqueName: \"kubernetes.io/projected/c6085b35-69c1-40e1-a782-d9f7f5708181-kube-api-access-xgllx\") pod \"cluster-samples-operator-665b6dd947-fs5p9\" (UID: \"c6085b35-69c1-40e1-a782-d9f7f5708181\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.789179 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.805966 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9q7w\" (UniqueName: \"kubernetes.io/projected/5d5efe87-5fb1-442e-b7bd-5c9758daa505-kube-api-access-t9q7w\") pod \"machine-approver-56656f9798-9d46g\" (UID: \"5d5efe87-5fb1-442e-b7bd-5c9758daa505\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.818056 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.822417 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjt7\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-kube-api-access-lmjt7\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.825573 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.844583 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xshqs\" (UniqueName: \"kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs\") pod \"oauth-openshift-558db77b4-9f545\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.847312 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.850768 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.890183 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ktd2\" (UniqueName: \"kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2\") pod \"controller-manager-879f6c89f-m6kfd\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.919933 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnpft\" (UniqueName: \"kubernetes.io/projected/3c1531f8-6f83-4253-a88f-b2ec8144f50d-kube-api-access-gnpft\") pod \"etcd-operator-b45778765-mlpbj\" (UID: \"3c1531f8-6f83-4253-a88f-b2ec8144f50d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.923592 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxwcp\" (UniqueName: \"kubernetes.io/projected/18b36f3e-aaad-4381-bf0f-af34a2c10337-kube-api-access-nxwcp\") pod \"openshift-controller-manager-operator-756b6f6bc6-4vw9n\" (UID: \"18b36f3e-aaad-4381-bf0f-af34a2c10337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.927831 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.945170 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.949021 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.961734 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.966442 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/79293b7f-3383-4376-90ca-1f6c6eff2dc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.969331 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 28 10:37:41 crc kubenswrapper[4972]: I0228 10:37:41.989994 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.059215 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j82kg\" (UniqueName: \"kubernetes.io/projected/c9173e02-eddc-40e3-86ab-0f0e8a8f50ac-kube-api-access-j82kg\") pod \"machine-api-operator-5694c8668f-7xwmd\" (UID: \"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.060449 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb996\" (UniqueName: \"kubernetes.io/projected/b0f314b2-1a1a-4235-841d-28872c4907ba-kube-api-access-vb996\") pod \"authentication-operator-69f744f599-jp6lb\" (UID: \"b0f314b2-1a1a-4235-841d-28872c4907ba\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.066133 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w8sp\" (UniqueName: \"kubernetes.io/projected/fcd06083-d0dc-4523-b832-471ca8b45e82-kube-api-access-8w8sp\") pod \"dns-operator-744455d44c-7gn5m\" (UID: \"fcd06083-d0dc-4523-b832-471ca8b45e82\") " pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.083951 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6bld\" (UniqueName: \"kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld\") pod \"console-f9d7485db-czzdx\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.096450 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.112809 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7dr\" (UniqueName: \"kubernetes.io/projected/398facc3-3546-42c1-8aac-07a59c92852c-kube-api-access-tf7dr\") pod \"console-operator-58897d9998-5d29s\" (UID: \"398facc3-3546-42c1-8aac-07a59c92852c\") " pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.113691 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.114186 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.135880 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.153732 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.165221 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kjqt\" (UniqueName: \"kubernetes.io/projected/9b27a0f8-30a8-4798-b7e7-612aa8313c2d-kube-api-access-7kjqt\") pod \"apiserver-7bbb656c7d-6s2p6\" (UID: \"9b27a0f8-30a8-4798-b7e7-612aa8313c2d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.169848 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.189326 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.190545 4972 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.205607 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.210681 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.223694 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.226231 4972 request.go:700] Waited for 1.919845881s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/configmaps?fieldSelector=metadata.name%3Dcni-sysctl-allowlist&limit=500&resourceVersion=0 Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.227807 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.245922 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.247834 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.259508 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.267955 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.269840 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/964ea0e3-fd84-4e16-873b-b3013209243d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-84psv\" (UID: \"964ea0e3-fd84-4e16-873b-b3013209243d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.289249 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.310071 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.312963 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.329200 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.348529 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.367987 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.392293 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.408883 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.428232 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.442037 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" event={"ID":"5d5efe87-5fb1-442e-b7bd-5c9758daa505","Type":"ContainerStarted","Data":"aefc039e1a62c00aed67e5c92472cc2a5f03fd227c38b491a480093cdbdc6eca"} Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.462249 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.469215 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl659\" (UniqueName: \"kubernetes.io/projected/40d5e1c5-21a4-419a-9425-6773e3f33c29-kube-api-access-gl659\") pod \"apiserver-76f77b778f-2xkbh\" (UID: \"40d5e1c5-21a4-419a-9425-6773e3f33c29\") " pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.485819 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69btw\" (UniqueName: \"kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw\") pod \"route-controller-manager-6576b87f9c-525bc\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.502915 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bj6\" (UniqueName: \"kubernetes.io/projected/eeb9fc08-9a96-445c-9e73-dd9a2b796d17-kube-api-access-68bj6\") pod \"service-ca-operator-777779d784-xstmm\" (UID: \"eeb9fc08-9a96-445c-9e73-dd9a2b796d17\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.513197 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.524434 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mfrtc"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.558303 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b637c49c-9ebf-40bc-8208-b3da5a05b4be-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x7r78\" (UID: \"b637c49c-9ebf-40bc-8208-b3da5a05b4be\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.570627 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgzgn\" (UniqueName: \"kubernetes.io/projected/8b782f84-114b-40d3-b7ba-806e1fccc04f-kube-api-access-xgzgn\") pod \"router-default-5444994796-4ftt6\" (UID: \"8b782f84-114b-40d3-b7ba-806e1fccc04f\") " pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.594791 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdml8\" (UniqueName: \"kubernetes.io/projected/cc45436c-617d-4941-a3ab-952c1bfd4d4f-kube-api-access-pdml8\") pod \"migrator-59844c95c7-bs4ks\" (UID: \"cc45436c-617d-4941-a3ab-952c1bfd4d4f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.607215 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnpd\" (UniqueName: \"kubernetes.io/projected/0389a687-8bad-46df-ac25-895f961d0184-kube-api-access-grnpd\") pod \"service-ca-9c57cc56f-599cv\" (UID: \"0389a687-8bad-46df-ac25-895f961d0184\") " pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.611688 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bd989"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.642878 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zchf\" (UniqueName: \"kubernetes.io/projected/63b91360-360a-46bf-ab58-6124f4210b84-kube-api-access-4zchf\") pod \"machine-config-controller-84d6567774-z2ghv\" (UID: \"63b91360-360a-46bf-ab58-6124f4210b84\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.666137 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wddsg\" (UniqueName: \"kubernetes.io/projected/6ba51b61-6394-4c9d-a5fb-389b2420f6e0-kube-api-access-wddsg\") pod \"machine-config-operator-74547568cd-hhfd6\" (UID: \"6ba51b61-6394-4c9d-a5fb-389b2420f6e0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.674553 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5d29s"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.679731 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a03f59e-a71b-49d2-9a93-59075788fc21-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmrtt\" (UID: \"2a03f59e-a71b-49d2-9a93-59075788fc21\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.683588 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnjz7\" (UniqueName: \"kubernetes.io/projected/06ae4169-c713-4ebc-88aa-dee750386952-kube-api-access-gnjz7\") pod \"olm-operator-6b444d44fb-p5xl7\" (UID: \"06ae4169-c713-4ebc-88aa-dee750386952\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:42 crc kubenswrapper[4972]: W0228 10:37:42.686644 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bb2eeb1_2033_4895_af78_9cf70134182e.slice/crio-cb6db013a8ab00d18134ee5adbe04271fb52c76f92e51d46a151a440a06d6138 WatchSource:0}: Error finding container cb6db013a8ab00d18134ee5adbe04271fb52c76f92e51d46a151a440a06d6138: Status 404 returned error can't find the container with id cb6db013a8ab00d18134ee5adbe04271fb52c76f92e51d46a151a440a06d6138 Feb 28 10:37:42 crc kubenswrapper[4972]: W0228 10:37:42.687813 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod398facc3_3546_42c1_8aac_07a59c92852c.slice/crio-1a2ef64fc374c491e616e59b84347b950ea9076999af4ee90f7a9de660d3c235 WatchSource:0}: Error finding container 1a2ef64fc374c491e616e59b84347b950ea9076999af4ee90f7a9de660d3c235: Status 404 returned error can't find the container with id 1a2ef64fc374c491e616e59b84347b950ea9076999af4ee90f7a9de660d3c235 Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.702816 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jp6lb"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.709351 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7k4\" (UniqueName: \"kubernetes.io/projected/79293b7f-3383-4376-90ca-1f6c6eff2dc4-kube-api-access-5f7k4\") pod \"package-server-manager-789f6589d5-2fwv6\" (UID: \"79293b7f-3383-4376-90ca-1f6c6eff2dc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.714872 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.715742 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9"] Feb 28 10:37:42 crc kubenswrapper[4972]: W0228 10:37:42.716217 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0f314b2_1a1a_4235_841d_28872c4907ba.slice/crio-4bf29ec2facb19f62207b482d903ecb05e3200b2871457f73b166c304f3a6996 WatchSource:0}: Error finding container 4bf29ec2facb19f62207b482d903ecb05e3200b2871457f73b166c304f3a6996: Status 404 returned error can't find the container with id 4bf29ec2facb19f62207b482d903ecb05e3200b2871457f73b166c304f3a6996 Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.726595 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.727944 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zwgg\" (UniqueName: \"kubernetes.io/projected/043d7cc4-5732-49b8-8150-bb897c5fa424-kube-api-access-7zwgg\") pod \"multus-admission-controller-857f4d67dd-44c45\" (UID: \"043d7cc4-5732-49b8-8150-bb897c5fa424\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.728960 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9f545"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.746310 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmhbq\" (UniqueName: \"kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq\") pod \"marketplace-operator-79b997595-c6hjj\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.748032 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.758447 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.764910 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.767966 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.772008 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7gn5m"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.774387 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.781398 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-599cv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.787594 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.791351 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.804787 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.807515 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.814193 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.814718 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.815652 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.827606 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 28 10:37:42 crc kubenswrapper[4972]: W0228 10:37:42.830023 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod964ea0e3_fd84_4e16_873b_b3013209243d.slice/crio-868bf5209e09597e111b2573ea24b2276bc251a2a30b998fdacf7cf3b876d75d WatchSource:0}: Error finding container 868bf5209e09597e111b2573ea24b2276bc251a2a30b998fdacf7cf3b876d75d: Status 404 returned error can't find the container with id 868bf5209e09597e111b2573ea24b2276bc251a2a30b998fdacf7cf3b876d75d Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.835401 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.842407 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.849135 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.856452 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.869946 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.884365 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.896118 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.908501 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xwmd"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.910441 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpbj"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.911940 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973697 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973754 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz2nb\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973780 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-srv-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973843 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973897 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-apiservice-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973932 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2btj7\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-kube-api-access-2btj7\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.973991 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxl7x\" (UniqueName: \"kubernetes.io/projected/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-kube-api-access-qxl7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974035 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b49304bc-f264-41e3-984a-baa952c23d12-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974056 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b49304bc-f264-41e3-984a-baa952c23d12-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974075 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974101 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b49304bc-f264-41e3-984a-baa952c23d12-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974124 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974200 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974227 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974261 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-webhook-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974286 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-profile-collector-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974315 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974337 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfc87\" (UniqueName: \"kubernetes.io/projected/121139d5-7a49-4920-ac95-fc77bfe36b58-kube-api-access-cfc87\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974373 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jxb\" (UniqueName: \"kubernetes.io/projected/6bb9bb11-63e1-4af7-bef6-71f60cf33766-kube-api-access-f9jxb\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974395 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974416 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974532 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974558 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974584 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974649 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974703 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j59g2\" (UniqueName: \"kubernetes.io/projected/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-kube-api-access-j59g2\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974792 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974819 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/121139d5-7a49-4920-ac95-fc77bfe36b58-tmpfs\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974844 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.974881 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgbzv\" (UniqueName: \"kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:42 crc kubenswrapper[4972]: E0228 10:37:42.976548 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.476529922 +0000 UTC m=+120.388537870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:42 crc kubenswrapper[4972]: I0228 10:37:42.983964 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.064673 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-599cv"] Feb 28 10:37:43 crc kubenswrapper[4972]: W0228 10:37:43.069603 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c1531f8_6f83_4253_a88f_b2ec8144f50d.slice/crio-7f675d587a0f88d546f5f1b5e5f12febfacd1930daadbd549808b1d48ebab86d WatchSource:0}: Error finding container 7f675d587a0f88d546f5f1b5e5f12febfacd1930daadbd549808b1d48ebab86d: Status 404 returned error can't find the container with id 7f675d587a0f88d546f5f1b5e5f12febfacd1930daadbd549808b1d48ebab86d Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075563 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075764 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-node-bootstrap-token\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075815 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.075833 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.575809478 +0000 UTC m=+120.487817216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075867 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075940 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.075992 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076059 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j59g2\" (UniqueName: \"kubernetes.io/projected/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-kube-api-access-j59g2\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076081 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-socket-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076156 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-registration-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076173 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgq2h\" (UniqueName: \"kubernetes.io/projected/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-kube-api-access-cgq2h\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076191 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6cnj\" (UniqueName: \"kubernetes.io/projected/7b440217-3770-4ec9-b252-868fe01353fa-kube-api-access-n6cnj\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076265 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076282 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/121139d5-7a49-4920-ac95-fc77bfe36b58-tmpfs\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076300 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076348 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4089e4d-bcd6-4524-bd70-b8eaf3768056-config-volume\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076364 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjv7p\" (UniqueName: \"kubernetes.io/projected/b4089e4d-bcd6-4524-bd70-b8eaf3768056-kube-api-access-wjv7p\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076381 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgbzv\" (UniqueName: \"kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076399 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076448 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qdjk\" (UniqueName: \"kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076480 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprmz\" (UniqueName: \"kubernetes.io/projected/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-kube-api-access-rprmz\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076518 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076537 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz2nb\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076566 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-srv-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076582 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076701 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-apiservice-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076719 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-cert\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076756 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-csi-data-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076782 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2btj7\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-kube-api-access-2btj7\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076812 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxl7x\" (UniqueName: \"kubernetes.io/projected/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-kube-api-access-qxl7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076826 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076841 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-mountpoint-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076857 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-plugins-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076896 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b49304bc-f264-41e3-984a-baa952c23d12-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076914 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b49304bc-f264-41e3-984a-baa952c23d12-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076928 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076964 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b49304bc-f264-41e3-984a-baa952c23d12-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076979 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.076998 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-certs\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077028 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4089e4d-bcd6-4524-bd70-b8eaf3768056-metrics-tls\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077053 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077071 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077087 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-webhook-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-profile-collector-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077172 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077188 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfc87\" (UniqueName: \"kubernetes.io/projected/121139d5-7a49-4920-ac95-fc77bfe36b58-kube-api-access-cfc87\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077255 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jxb\" (UniqueName: \"kubernetes.io/projected/6bb9bb11-63e1-4af7-bef6-71f60cf33766-kube-api-access-f9jxb\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077276 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077291 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.077359 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.079147 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.083081 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.583067052 +0000 UTC m=+120.495074790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.083403 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/121139d5-7a49-4920-ac95-fc77bfe36b58-tmpfs\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.083853 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.089412 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.090424 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.095251 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-apiservice-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.096968 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.097264 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.097514 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b49304bc-f264-41e3-984a-baa952c23d12-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.097861 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/121139d5-7a49-4920-ac95-fc77bfe36b58-webhook-cert\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.098707 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.104315 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.104727 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.104732 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.105250 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.105346 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.105669 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b49304bc-f264-41e3-984a-baa952c23d12-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.106995 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.109903 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-srv-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.110362 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6bb9bb11-63e1-4af7-bef6-71f60cf33766-profile-collector-cert\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.111161 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.127557 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.136760 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2xkbh"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.147933 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j59g2\" (UniqueName: \"kubernetes.io/projected/f7d82d58-d793-4201-b1c0-4ad0b6c28a20-kube-api-access-j59g2\") pod \"kube-storage-version-migrator-operator-b67b599dd-gmfff\" (UID: \"f7d82d58-d793-4201-b1c0-4ad0b6c28a20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.161302 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgbzv\" (UniqueName: \"kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv\") pod \"collect-profiles-29537910-gdj48\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183224 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183408 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-socket-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183439 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-registration-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183609 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgq2h\" (UniqueName: \"kubernetes.io/projected/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-kube-api-access-cgq2h\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183635 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6cnj\" (UniqueName: \"kubernetes.io/projected/7b440217-3770-4ec9-b252-868fe01353fa-kube-api-access-n6cnj\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183675 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4089e4d-bcd6-4524-bd70-b8eaf3768056-config-volume\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183693 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjv7p\" (UniqueName: \"kubernetes.io/projected/b4089e4d-bcd6-4524-bd70-b8eaf3768056-kube-api-access-wjv7p\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183712 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183733 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qdjk\" (UniqueName: \"kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183771 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprmz\" (UniqueName: \"kubernetes.io/projected/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-kube-api-access-rprmz\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183827 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-cert\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183851 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-csi-data-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183911 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183933 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-mountpoint-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183955 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-plugins-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.183990 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-certs\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.184016 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4089e4d-bcd6-4524-bd70-b8eaf3768056-metrics-tls\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.184076 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.184117 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-node-bootstrap-token\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.184943 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.684922204 +0000 UTC m=+120.596929942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.185313 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-socket-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.186395 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-registration-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.186564 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-mountpoint-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.187158 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.187188 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-csi-data-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.187889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-node-bootstrap-token\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.187945 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b440217-3770-4ec9-b252-868fe01353fa-plugins-dir\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.188061 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4089e4d-bcd6-4524-bd70-b8eaf3768056-config-volume\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.189233 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.189420 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.190932 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-certs\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.191018 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-cert\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.191594 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4089e4d-bcd6-4524-bd70-b8eaf3768056-metrics-tls\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.199093 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b49304bc-f264-41e3-984a-baa952c23d12-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bdb62\" (UID: \"b49304bc-f264-41e3-984a-baa952c23d12\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: W0228 10:37:43.208228 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0389a687_8bad_46df_ac25_895f961d0184.slice/crio-d47b1156a62c23876801cd3455eed031bff9ae99a5e1814131c70170f9b0b128 WatchSource:0}: Error finding container d47b1156a62c23876801cd3455eed031bff9ae99a5e1814131c70170f9b0b128: Status 404 returned error can't find the container with id d47b1156a62c23876801cd3455eed031bff9ae99a5e1814131c70170f9b0b128 Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.220479 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.229231 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz2nb\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.243015 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.250760 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2btj7\" (UniqueName: \"kubernetes.io/projected/ccfa71dd-d16e-482e-88f0-2b1a634f44b6-kube-api-access-2btj7\") pod \"ingress-operator-5b745b69d9-vjxlj\" (UID: \"ccfa71dd-d16e-482e-88f0-2b1a634f44b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.269573 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxl7x\" (UniqueName: \"kubernetes.io/projected/8066e21c-fa0c-497d-8bc8-a27d07dab5b4-kube-api-access-qxl7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-gzmqk\" (UID: \"8066e21c-fa0c-497d-8bc8-a27d07dab5b4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.284669 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfc87\" (UniqueName: \"kubernetes.io/projected/121139d5-7a49-4920-ac95-fc77bfe36b58-kube-api-access-cfc87\") pod \"packageserver-d55dfcdfc-kcr2z\" (UID: \"121139d5-7a49-4920-ac95-fc77bfe36b58\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.291276 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.291860 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.791836786 +0000 UTC m=+120.703844694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.309204 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jxb\" (UniqueName: \"kubernetes.io/projected/6bb9bb11-63e1-4af7-bef6-71f60cf33766-kube-api-access-f9jxb\") pod \"catalog-operator-68c6474976-wntf7\" (UID: \"6bb9bb11-63e1-4af7-bef6-71f60cf33766\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.329419 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qdjk\" (UniqueName: \"kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk\") pod \"cni-sysctl-allowlist-ds-5plg9\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.340121 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.347648 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6cnj\" (UniqueName: \"kubernetes.io/projected/7b440217-3770-4ec9-b252-868fe01353fa-kube-api-access-n6cnj\") pod \"csi-hostpathplugin-jnb9w\" (UID: \"7b440217-3770-4ec9-b252-868fe01353fa\") " pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.373223 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgq2h\" (UniqueName: \"kubernetes.io/projected/598063bd-cd95-485e-ab4b-ca1d3cc7a4ff-kube-api-access-cgq2h\") pod \"machine-config-server-nfhbz\" (UID: \"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff\") " pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.385946 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprmz\" (UniqueName: \"kubernetes.io/projected/699276a9-dfad-4ed8-9dcf-e197aa1d0a10-kube-api-access-rprmz\") pod \"ingress-canary-nkn8q\" (UID: \"699276a9-dfad-4ed8-9dcf-e197aa1d0a10\") " pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.395056 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.895021767 +0000 UTC m=+120.807029525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.396188 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.396681 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.397045 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.897032316 +0000 UTC m=+120.809040054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.417129 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjv7p\" (UniqueName: \"kubernetes.io/projected/b4089e4d-bcd6-4524-bd70-b8eaf3768056-kube-api-access-wjv7p\") pod \"dns-default-zxcst\" (UID: \"b4089e4d-bcd6-4524-bd70-b8eaf3768056\") " pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.418706 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.425862 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.450117 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78"] Feb 28 10:37:43 crc kubenswrapper[4972]: W0228 10:37:43.463422 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a03f59e_a71b_49d2_9a93_59075788fc21.slice/crio-c18f67800df5358bb9b037f0df9510f9b46680959d53b45295e247e575a1b5fb WatchSource:0}: Error finding container c18f67800df5358bb9b037f0df9510f9b46680959d53b45295e247e575a1b5fb: Status 404 returned error can't find the container with id c18f67800df5358bb9b037f0df9510f9b46680959d53b45295e247e575a1b5fb Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.478382 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.497769 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.498189 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:43.998153556 +0000 UTC m=+120.910161394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.514849 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.518655 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.532035 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.542324 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-599cv" event={"ID":"0389a687-8bad-46df-ac25-895f961d0184","Type":"ContainerStarted","Data":"d47b1156a62c23876801cd3455eed031bff9ae99a5e1814131c70170f9b0b128"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.567525 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.575215 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xstmm"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.575807 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" event={"ID":"5da2e110-e194-4b21-a995-e7fcd1bb1825","Type":"ContainerStarted","Data":"e82a11a53798b6c922b808cb6ed6033a56bf3755640ac5d0e695a15245956f54"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.579538 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.587709 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mfrtc" event={"ID":"3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53","Type":"ContainerStarted","Data":"de4d9348ac1f0f4636badd0e093eaaa4c84b6b792e3b76d6928dfa6629e1c2aa"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.587764 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mfrtc" event={"ID":"3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53","Type":"ContainerStarted","Data":"488176d4e188f2169ece03806a2c57de125c66c34ed77cf70272b9597e06abe1"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.588791 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.588798 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nkn8q" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.592065 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4ftt6" event={"ID":"8b782f84-114b-40d3-b7ba-806e1fccc04f","Type":"ContainerStarted","Data":"504102495e81c9c59ebaa8a574ba130a33fb8ba3f94f1cf138d258cb87331776"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.593994 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5d29s" event={"ID":"398facc3-3546-42c1-8aac-07a59c92852c","Type":"ContainerStarted","Data":"2b2cecf322a0a2fe5019637b46cdcaa0d89d6d4564ef11670633ad4f117f55be"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.594025 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5d29s" event={"ID":"398facc3-3546-42c1-8aac-07a59c92852c","Type":"ContainerStarted","Data":"1a2ef64fc374c491e616e59b84347b950ea9076999af4ee90f7a9de660d3c235"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.594398 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.594443 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.600719 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-nfhbz" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.601164 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.601604 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.101588254 +0000 UTC m=+121.013595992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.608687 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" event={"ID":"9bb2eeb1-2033-4895-af78-9cf70134182e","Type":"ContainerStarted","Data":"90776a6db2d66079ac9625f93ef695f28687935c258817650180ae2563420783"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.608752 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" event={"ID":"9bb2eeb1-2033-4895-af78-9cf70134182e","Type":"ContainerStarted","Data":"cb6db013a8ab00d18134ee5adbe04271fb52c76f92e51d46a151a440a06d6138"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.614853 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" event={"ID":"3c1531f8-6f83-4253-a88f-b2ec8144f50d","Type":"ContainerStarted","Data":"7f675d587a0f88d546f5f1b5e5f12febfacd1930daadbd549808b1d48ebab86d"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.617369 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" event={"ID":"0d27c7ba-b3c5-492a-9a57-53a179cc40d9","Type":"ContainerStarted","Data":"3ded79de9651c6d1cebe41040b6cf621b6e1990131701708a1ad25a7627603eb"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.617789 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.618391 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.621083 4972 patch_prober.go:28] interesting pod/console-operator-58897d9998-5d29s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.621154 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5d29s" podUID="398facc3-3546-42c1-8aac-07a59c92852c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.621216 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-czzdx" event={"ID":"6505a931-94e8-45b7-8945-61f412bd5fd1","Type":"ContainerStarted","Data":"4cd378189333db82536e0be7c690cbcfbbdcca29d1b902bc87d9d5c0999e7dd6"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.654057 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" event={"ID":"b0f314b2-1a1a-4235-841d-28872c4907ba","Type":"ContainerStarted","Data":"4bf29ec2facb19f62207b482d903ecb05e3200b2871457f73b166c304f3a6996"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.662209 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" event={"ID":"9b27a0f8-30a8-4798-b7e7-612aa8313c2d","Type":"ContainerStarted","Data":"4ba47404c13a957720bacd9d911cacdb800940d9743ef0258ef836e9db5f6a75"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.670210 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.675296 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.686744 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" event={"ID":"8b62a70e-5da5-48a8-9eae-e5b529606f3c","Type":"ContainerStarted","Data":"c4f64e3fbdb115ea149395d9d80426ae40da66b8a82f674880b8cdf645be55e0"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.686784 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" event={"ID":"8b62a70e-5da5-48a8-9eae-e5b529606f3c","Type":"ContainerStarted","Data":"a1a141c13799c439c4f813bca7afa7a83ccb846e053068f38be91b8d5497e624"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.694966 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" event={"ID":"63f17722-9aca-469f-85f4-3cb6774a90ea","Type":"ContainerStarted","Data":"1df56dd5d529843394c853ca8041f9d9b7c9ca6c2d165d5761829ca551133984"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.701794 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.701934 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.20191649 +0000 UTC m=+121.113924228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.702185 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.702583 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.202564009 +0000 UTC m=+121.114571807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.710006 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" event={"ID":"5d5efe87-5fb1-442e-b7bd-5c9758daa505","Type":"ContainerStarted","Data":"40876298b737707227e70f6401b2b6e1be2eca1c5469b000b881e66292a2a38f"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.711722 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" event={"ID":"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac","Type":"ContainerStarted","Data":"94e17fd4d7071f0b7f4c77226fb3be31637f36989dcc69cc7bf1f82edfe4e7e4"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.715691 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" event={"ID":"964ea0e3-fd84-4e16-873b-b3013209243d","Type":"ContainerStarted","Data":"868bf5209e09597e111b2573ea24b2276bc251a2a30b998fdacf7cf3b876d75d"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.718569 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" event={"ID":"40d5e1c5-21a4-419a-9425-6773e3f33c29","Type":"ContainerStarted","Data":"133c2449f37788cd5a5d5570f063c6173e031ea232271d7d537efe8f15cc0f14"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.720060 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" event={"ID":"c6085b35-69c1-40e1-a782-d9f7f5708181","Type":"ContainerStarted","Data":"447b93f0f9ee29e1c8dccb637d64b28658c19506becbbe8a3b506fc2ce9098cf"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.727796 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" event={"ID":"fcd06083-d0dc-4523-b832-471ca8b45e82","Type":"ContainerStarted","Data":"149294dc889fb4b7f989236b2ba5b7027e4e7e6c4de69bf554d759a0e6ae1cd6"} Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.732611 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" event={"ID":"18b36f3e-aaad-4381-bf0f-af34a2c10337","Type":"ContainerStarted","Data":"1ba34d8c788ed480d794199d082be88f69a3d6ec3bba624edfa783ea6fe3ef03"} Feb 28 10:37:43 crc kubenswrapper[4972]: W0228 10:37:43.735946 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb637c49c_9ebf_40bc_8208_b3da5a05b4be.slice/crio-464d78ed30bdb07d6d6e888ef0cdd57a4ab2e84fdb0f806f47abe90b19970c65 WatchSource:0}: Error finding container 464d78ed30bdb07d6d6e888ef0cdd57a4ab2e84fdb0f806f47abe90b19970c65: Status 404 returned error can't find the container with id 464d78ed30bdb07d6d6e888ef0cdd57a4ab2e84fdb0f806f47abe90b19970c65 Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.803194 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.803363 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.303333298 +0000 UTC m=+121.215341036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.803864 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.804236 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.304219035 +0000 UTC m=+121.216226853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: W0228 10:37:43.888835 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc45436c_617d_4941_a3ab_952c1bfd4d4f.slice/crio-293b0e9347acbb8f19a4412303e05dddc218900bc0b9ad203503dbc274ad8914 WatchSource:0}: Error finding container 293b0e9347acbb8f19a4412303e05dddc218900bc0b9ad203503dbc274ad8914: Status 404 returned error can't find the container with id 293b0e9347acbb8f19a4412303e05dddc218900bc0b9ad203503dbc274ad8914 Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.905258 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:43 crc kubenswrapper[4972]: E0228 10:37:43.908206 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.408166819 +0000 UTC m=+121.320174567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.940611 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" podStartSLOduration=51.940529458 podStartE2EDuration="51.940529458s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:43.899705038 +0000 UTC m=+120.811712786" watchObservedRunningTime="2026-02-28 10:37:43.940529458 +0000 UTC m=+120.852537196" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.941628 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.943158 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv"] Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.983030 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mfrtc" podStartSLOduration=51.983007419 podStartE2EDuration="51.983007419s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:43.95608958 +0000 UTC m=+120.868097318" watchObservedRunningTime="2026-02-28 10:37:43.983007419 +0000 UTC m=+120.895015157" Feb 28 10:37:43 crc kubenswrapper[4972]: I0228 10:37:43.990014 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.010507 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.010940 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.510914907 +0000 UTC m=+121.422922645 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.028440 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.048247 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-44c45"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.108691 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.114629 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.115052 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.615035886 +0000 UTC m=+121.527043624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.118805 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79293b7f_3383_4376_90ca_1f6c6eff2dc4.slice/crio-41c134e33117c013fb34169f8a239b6f2cffd0ee500007d8534b1a4c17d4740c WatchSource:0}: Error finding container 41c134e33117c013fb34169f8a239b6f2cffd0ee500007d8534b1a4c17d4740c: Status 404 returned error can't find the container with id 41c134e33117c013fb34169f8a239b6f2cffd0ee500007d8534b1a4c17d4740c Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.123982 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b91360_360a_46bf_ab58_6124f4210b84.slice/crio-76d783ca9dcf0331b46062c7c00e7433eec1bcd03b7210caf9d7ce5d618d2aee WatchSource:0}: Error finding container 76d783ca9dcf0331b46062c7c00e7433eec1bcd03b7210caf9d7ce5d618d2aee: Status 404 returned error can't find the container with id 76d783ca9dcf0331b46062c7c00e7433eec1bcd03b7210caf9d7ce5d618d2aee Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.133047 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod043d7cc4_5732_49b8_8150_bb897c5fa424.slice/crio-ed6aaa70bb9f00c8f56c2ca45cba2fbb12c46b8e4c60150a601f85d997a6f656 WatchSource:0}: Error finding container ed6aaa70bb9f00c8f56c2ca45cba2fbb12c46b8e4c60150a601f85d997a6f656: Status 404 returned error can't find the container with id ed6aaa70bb9f00c8f56c2ca45cba2fbb12c46b8e4c60150a601f85d997a6f656 Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.184647 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb49304bc_f264_41e3_984a_baa952c23d12.slice/crio-f9471efd0880c407fcb62f6a7dda59b64b74f3989f355835450faf68bb0d1a0a WatchSource:0}: Error finding container f9471efd0880c407fcb62f6a7dda59b64b74f3989f355835450faf68bb0d1a0a: Status 404 returned error can't find the container with id f9471efd0880c407fcb62f6a7dda59b64b74f3989f355835450faf68bb0d1a0a Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.215806 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.216109 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.716097774 +0000 UTC m=+121.628105512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.317191 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.318221 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.818192452 +0000 UTC m=+121.730200190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.319002 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.319285 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.819272584 +0000 UTC m=+121.731280322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.344855 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5d29s" podStartSLOduration=52.344840423 podStartE2EDuration="52.344840423s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:44.343864963 +0000 UTC m=+121.255872701" watchObservedRunningTime="2026-02-28 10:37:44.344840423 +0000 UTC m=+121.256848161" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.384347 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.424907 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.425156 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.925114643 +0000 UTC m=+121.837122381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.425287 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.425733 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:44.925725752 +0000 UTC m=+121.837733490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.431816 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jnb9w"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.478167 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.530622 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.531482 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.031441088 +0000 UTC m=+121.943448826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.545996 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z"] Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.561123 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccfa71dd_d16e_482e_88f0_2b1a634f44b6.slice/crio-65c98e710349e43e8607820e25b0e5ff3bdbc4d2100f2a821d5272634680626d WatchSource:0}: Error finding container 65c98e710349e43e8607820e25b0e5ff3bdbc4d2100f2a821d5272634680626d: Status 404 returned error can't find the container with id 65c98e710349e43e8607820e25b0e5ff3bdbc4d2100f2a821d5272634680626d Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.618714 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.629892 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nkn8q"] Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.632134 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.632573 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.132556077 +0000 UTC m=+122.044563815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: W0228 10:37:44.662416 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1efe1619_53ab_4d7b_83d8_c772875cf078.slice/crio-dc35327ec7148954bef1417a3e030fb2da3dfb3643bf129c2fbdd527261baf5e WatchSource:0}: Error finding container dc35327ec7148954bef1417a3e030fb2da3dfb3643bf129c2fbdd527261baf5e: Status 404 returned error can't find the container with id dc35327ec7148954bef1417a3e030fb2da3dfb3643bf129c2fbdd527261baf5e Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.679052 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lmtss" podStartSLOduration=52.679033456 podStartE2EDuration="52.679033456s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:44.677314615 +0000 UTC m=+121.589322353" watchObservedRunningTime="2026-02-28 10:37:44.679033456 +0000 UTC m=+121.591041194" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.733196 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.733366 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.733419 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.733444 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.733475 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.733641 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.233625606 +0000 UTC m=+122.145633344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.761363 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4ftt6" event={"ID":"8b782f84-114b-40d3-b7ba-806e1fccc04f","Type":"ContainerStarted","Data":"b07b6400b1f605c1355463bb2594d3f675daa0e1f27226205090ca0944c9d7d1"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.768565 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.773993 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" event={"ID":"79293b7f-3383-4376-90ca-1f6c6eff2dc4","Type":"ContainerStarted","Data":"41c134e33117c013fb34169f8a239b6f2cffd0ee500007d8534b1a4c17d4740c"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.777244 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" event={"ID":"06ae4169-c713-4ebc-88aa-dee750386952","Type":"ContainerStarted","Data":"0ba83e0f1ef836557310fefc3e95c46b6849c595fdc9bbdf52a2374ea78d8979"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.778335 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" event={"ID":"b0f314b2-1a1a-4235-841d-28872c4907ba","Type":"ContainerStarted","Data":"bf660134ff338728ac7871bdcc31aa6d9fe3120720971cabfdbcc770182292ce"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.782500 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.787686 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.791040 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" event={"ID":"043d7cc4-5732-49b8-8150-bb897c5fa424","Type":"ContainerStarted","Data":"ed6aaa70bb9f00c8f56c2ca45cba2fbb12c46b8e4c60150a601f85d997a6f656"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.805979 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.810252 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.810295 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.810367 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.828956 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" event={"ID":"fcd06083-d0dc-4523-b832-471ca8b45e82","Type":"ContainerStarted","Data":"3930befbd13edd1cb8f101d83ae95a3e12b09b67ac427978183377caa4f0780e"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.834378 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" event={"ID":"f7d82d58-d793-4201-b1c0-4ad0b6c28a20","Type":"ContainerStarted","Data":"3344a45c4aa0fb601d3b6bd78c2f951b05c7fac37036a1faf1dcb1493f72b78c"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.834810 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.835925 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.33590882 +0000 UTC m=+122.247916688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.837208 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" event={"ID":"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac","Type":"ContainerStarted","Data":"e9788aeb67b95e43ef9499fcc7e01213c4660388c29f0c660fe1a5aff0df8581"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.841664 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.842133 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" event={"ID":"ccfa71dd-d16e-482e-88f0-2b1a634f44b6","Type":"ContainerStarted","Data":"65c98e710349e43e8607820e25b0e5ff3bdbc4d2100f2a821d5272634680626d"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.845597 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" event={"ID":"63f17722-9aca-469f-85f4-3cb6774a90ea","Type":"ContainerStarted","Data":"c6ff45262550ae89793c07ef093f0db31732d7fe21982ace0ab045397591cab1"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.846664 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.847722 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.848737 4972 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9f545 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.848822 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.856000 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" event={"ID":"cc45436c-617d-4941-a3ab-952c1bfd4d4f","Type":"ContainerStarted","Data":"293b0e9347acbb8f19a4412303e05dddc218900bc0b9ad203503dbc274ad8914"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.860368 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.862024 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.880852 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" event={"ID":"18b36f3e-aaad-4381-bf0f-af34a2c10337","Type":"ContainerStarted","Data":"2abbf189207a49cd83e02ae5039664790ec0f6ff7ae51b2e87cef20a802cf48f"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.883864 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nfhbz" event={"ID":"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff","Type":"ContainerStarted","Data":"f7cb86579aa6ddd56a11c5aa091b3c8b18237a7edb2d3d9e61fce729c2b018a3"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.885426 4972 generic.go:334] "Generic (PLEG): container finished" podID="9bb2eeb1-2033-4895-af78-9cf70134182e" containerID="90776a6db2d66079ac9625f93ef695f28687935c258817650180ae2563420783" exitCode=0 Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.885503 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" event={"ID":"9bb2eeb1-2033-4895-af78-9cf70134182e","Type":"ContainerDied","Data":"90776a6db2d66079ac9625f93ef695f28687935c258817650180ae2563420783"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.887701 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" event={"ID":"b49304bc-f264-41e3-984a-baa952c23d12","Type":"ContainerStarted","Data":"f9471efd0880c407fcb62f6a7dda59b64b74f3989f355835450faf68bb0d1a0a"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.889290 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" event={"ID":"0d27c7ba-b3c5-492a-9a57-53a179cc40d9","Type":"ContainerStarted","Data":"69d364177737fc390b663d73b69d7071a89b5009461a712a071667ca821c8333"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.889552 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.890650 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" event={"ID":"7b440217-3770-4ec9-b252-868fe01353fa","Type":"ContainerStarted","Data":"7ca3f1e277ce26bf7295c10e3b2f20cd1d98b15e4fb7ff763eb620c108fd836d"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.891488 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" event={"ID":"0dbb4a63-be87-47b1-8650-59c40ed5f7b5","Type":"ContainerStarted","Data":"7b41de8b1b785cc99ce8e49a93ee815446c0e770fc1d1933c063f1eaf9265cd9"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.893353 4972 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-m6kfd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.893395 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.894582 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" event={"ID":"1efe1619-53ab-4d7b-83d8-c772875cf078","Type":"ContainerStarted","Data":"dc35327ec7148954bef1417a3e030fb2da3dfb3643bf129c2fbdd527261baf5e"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.898564 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" event={"ID":"5d5efe87-5fb1-442e-b7bd-5c9758daa505","Type":"ContainerStarted","Data":"71cdfbef2d6c0b2d80d8bedbed03f04d1eb04680df40b1ecb28240e973bc7117"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.900479 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" event={"ID":"121139d5-7a49-4920-ac95-fc77bfe36b58","Type":"ContainerStarted","Data":"a110da0686b94c0846426c38a067931fd175995fbb763457a9db1e01dd08d39a"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.902416 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" event={"ID":"6ba51b61-6394-4c9d-a5fb-389b2420f6e0","Type":"ContainerStarted","Data":"bf95013e3e4c9ae8cff3bffd8b615c311903efe96ad5fa44fe73f20bfc692364"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.903578 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" event={"ID":"8066e21c-fa0c-497d-8bc8-a27d07dab5b4","Type":"ContainerStarted","Data":"f46d2bfab3fe344dc138042f966c8037a80b30adb5e5e24df154c7794128d8b4"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.904523 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" event={"ID":"eeb9fc08-9a96-445c-9e73-dd9a2b796d17","Type":"ContainerStarted","Data":"8d211d4c51ec4bd3f6b80569e86c409a5482f03c22e19d93eefad47968f9af50"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.907927 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-84psv" event={"ID":"964ea0e3-fd84-4e16-873b-b3013209243d","Type":"ContainerStarted","Data":"a6370b30b72ec624b0f9e8363c13ede7fb5f3f706dac39d6f9603288f66926d9"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.909943 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" event={"ID":"63b91360-360a-46bf-ab58-6124f4210b84","Type":"ContainerStarted","Data":"76d783ca9dcf0331b46062c7c00e7433eec1bcd03b7210caf9d7ce5d618d2aee"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.912268 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-czzdx" event={"ID":"6505a931-94e8-45b7-8945-61f412bd5fd1","Type":"ContainerStarted","Data":"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.913580 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" event={"ID":"dde838bc-7c41-4da3-a00d-bb704e436faf","Type":"ContainerStarted","Data":"152ec1bca1ca80037a09b0f50cbf0de64cbd2e68fdbb09c2d9fbb37ec0e88c37"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.924530 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" event={"ID":"2a03f59e-a71b-49d2-9a93-59075788fc21","Type":"ContainerStarted","Data":"c18f67800df5358bb9b037f0df9510f9b46680959d53b45295e247e575a1b5fb"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.928240 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-599cv" event={"ID":"0389a687-8bad-46df-ac25-895f961d0184","Type":"ContainerStarted","Data":"2bb3e7416739a3fe686aa28dcb69a1ea71515b78a2d22dd793356edfc3b5ee3e"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.930150 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" event={"ID":"b637c49c-9ebf-40bc-8208-b3da5a05b4be","Type":"ContainerStarted","Data":"464d78ed30bdb07d6d6e888ef0cdd57a4ab2e84fdb0f806f47abe90b19970c65"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.932342 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" event={"ID":"c6085b35-69c1-40e1-a782-d9f7f5708181","Type":"ContainerStarted","Data":"61d631026ab98be0a0e4b5c48a06344558427f7b4c074d24ad71204355b55553"} Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.933014 4972 patch_prober.go:28] interesting pod/console-operator-58897d9998-5d29s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.933072 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5d29s" podUID="398facc3-3546-42c1-8aac-07a59c92852c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.933682 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.933715 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.936495 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.936706 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.436676759 +0000 UTC m=+122.348684497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.936796 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.936952 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:44 crc kubenswrapper[4972]: E0228 10:37:44.939345 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.439323988 +0000 UTC m=+122.351331726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.970843 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.983489 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7f1c1c7-f8f6-418e-8526-e7946891f02e-metrics-certs\") pod \"network-metrics-daemon-lcv58\" (UID: \"b7f1c1c7-f8f6-418e-8526-e7946891f02e\") " pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:44 crc kubenswrapper[4972]: I0228 10:37:44.995505 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7"] Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.000780 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.008549 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.024270 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zxcst"] Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.029248 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.037601 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lcv58" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.038563 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.040153 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.540136328 +0000 UTC m=+122.452144066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.076709 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4vw9n" podStartSLOduration=53.076685782 podStartE2EDuration="53.076685782s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.074756905 +0000 UTC m=+121.986764643" watchObservedRunningTime="2026-02-28 10:37:45.076685782 +0000 UTC m=+121.988693510" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.102015 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.140249 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.140717 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.640703232 +0000 UTC m=+122.552710970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: W0228 10:37:45.237188 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bb9bb11_63e1_4af7_bef6_71f60cf33766.slice/crio-1022a9da89179e9a707b3656d1b083a437dd084c4f455d432e682f7075e3126c WatchSource:0}: Error finding container 1022a9da89179e9a707b3656d1b083a437dd084c4f455d432e682f7075e3126c: Status 404 returned error can't find the container with id 1022a9da89179e9a707b3656d1b083a437dd084c4f455d432e682f7075e3126c Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.241434 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.241572 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.741546703 +0000 UTC m=+122.653554441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.241816 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.242208 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.742199922 +0000 UTC m=+122.654207660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.322699 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" podStartSLOduration=53.322662159 podStartE2EDuration="53.322662159s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.322074442 +0000 UTC m=+122.234082180" watchObservedRunningTime="2026-02-28 10:37:45.322662159 +0000 UTC m=+122.234669897" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.324755 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" podStartSLOduration=53.324746491 podStartE2EDuration="53.324746491s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.281962212 +0000 UTC m=+122.193969950" watchObservedRunningTime="2026-02-28 10:37:45.324746491 +0000 UTC m=+122.236754249" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.362372 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.363213 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.863168931 +0000 UTC m=+122.775176669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.381080 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jp6lb" podStartSLOduration=53.381054452 podStartE2EDuration="53.381054452s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.37394797 +0000 UTC m=+122.285955708" watchObservedRunningTime="2026-02-28 10:37:45.381054452 +0000 UTC m=+122.293062190" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.412283 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-4ftt6" podStartSLOduration=53.412254247 podStartE2EDuration="53.412254247s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.401389335 +0000 UTC m=+122.313397103" watchObservedRunningTime="2026-02-28 10:37:45.412254247 +0000 UTC m=+122.324261995" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.445873 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9d46g" podStartSLOduration=53.445843874 podStartE2EDuration="53.445843874s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.439626609 +0000 UTC m=+122.351634347" watchObservedRunningTime="2026-02-28 10:37:45.445843874 +0000 UTC m=+122.357851612" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.465119 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.472036 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:45.971998219 +0000 UTC m=+122.884005957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.512406 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-czzdx" podStartSLOduration=53.512387688 podStartE2EDuration="53.512387688s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.510009467 +0000 UTC m=+122.422017205" watchObservedRunningTime="2026-02-28 10:37:45.512387688 +0000 UTC m=+122.424395426" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.571092 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.571287 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.071253874 +0000 UTC m=+122.983261612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.572158 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.573005 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.072988435 +0000 UTC m=+122.984996173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.621008 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-599cv" podStartSLOduration=53.620987039 podStartE2EDuration="53.620987039s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.546136069 +0000 UTC m=+122.458143807" watchObservedRunningTime="2026-02-28 10:37:45.620987039 +0000 UTC m=+122.532994787" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.675053 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.675266 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.175240279 +0000 UTC m=+123.087248017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.675444 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.675932 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.175924899 +0000 UTC m=+123.087932637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.722702 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lcv58"] Feb 28 10:37:45 crc kubenswrapper[4972]: W0228 10:37:45.773912 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7f1c1c7_f8f6_418e_8526_e7946891f02e.slice/crio-1183d401c097fdc558834b64afb61c1b76ccd28f49c5aefb5275609074f1f6e9 WatchSource:0}: Error finding container 1183d401c097fdc558834b64afb61c1b76ccd28f49c5aefb5275609074f1f6e9: Status 404 returned error can't find the container with id 1183d401c097fdc558834b64afb61c1b76ccd28f49c5aefb5275609074f1f6e9 Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.776725 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.776964 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.276948306 +0000 UTC m=+123.188956044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.807811 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.807869 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.877415 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.877755 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.377742586 +0000 UTC m=+123.289750324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.939303 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" event={"ID":"6ba51b61-6394-4c9d-a5fb-389b2420f6e0","Type":"ContainerStarted","Data":"5a086eb8ab790cafbde6153fca2820db15f0a9c7ccfa0d492300f150c42d73b9"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.940614 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lcv58" event={"ID":"b7f1c1c7-f8f6-418e-8526-e7946891f02e","Type":"ContainerStarted","Data":"1183d401c097fdc558834b64afb61c1b76ccd28f49c5aefb5275609074f1f6e9"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.942432 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-nfhbz" event={"ID":"598063bd-cd95-485e-ab4b-ca1d3cc7a4ff","Type":"ContainerStarted","Data":"5501b5e6c04206c715d11414ac66ed299c559a5986c1721e5843380eb8f6df42"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.943656 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" event={"ID":"b49304bc-f264-41e3-984a-baa952c23d12","Type":"ContainerStarted","Data":"6f1d7738dc6242d0e0f7042c0999004a3920dc899c80f450b89b1da18fee2531"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.944649 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" event={"ID":"2a03f59e-a71b-49d2-9a93-59075788fc21","Type":"ContainerStarted","Data":"14e4ea8cc419e172201d46dd4d1d16a51bbcb06572f1da6825ea107a775c685b"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.949605 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nkn8q" event={"ID":"699276a9-dfad-4ed8-9dcf-e197aa1d0a10","Type":"ContainerStarted","Data":"70fe09bc6d6f3c9f95c7579c7f486d628b3b53baf4ab560046a68f9ffd2c9d32"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.950944 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" event={"ID":"6bb9bb11-63e1-4af7-bef6-71f60cf33766","Type":"ContainerStarted","Data":"1022a9da89179e9a707b3656d1b083a437dd084c4f455d432e682f7075e3126c"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.952115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" event={"ID":"79293b7f-3383-4376-90ca-1f6c6eff2dc4","Type":"ContainerStarted","Data":"8dab834c6a5818bc5f1d673031c1b85f678aa19b37af9e08b5783ecce3bb5974"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.953996 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" event={"ID":"b637c49c-9ebf-40bc-8208-b3da5a05b4be","Type":"ContainerStarted","Data":"069dccb92aa3816c6819abb074a0fed02f65a895398b1e8547c830660acbfc1a"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.958787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zxcst" event={"ID":"b4089e4d-bcd6-4524-bd70-b8eaf3768056","Type":"ContainerStarted","Data":"0ad9c6351b45fb327143a010928be1d9ad135e975748e34c88c01974522ff36f"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.960523 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmrtt" podStartSLOduration=53.96051138 podStartE2EDuration="53.96051138s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.959116559 +0000 UTC m=+122.871124307" watchObservedRunningTime="2026-02-28 10:37:45.96051138 +0000 UTC m=+122.872519138" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.962097 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" event={"ID":"043d7cc4-5732-49b8-8150-bb897c5fa424","Type":"ContainerStarted","Data":"5c9a4c9dbac0318df7a8d701403361b8d8a58a12453538f7e96f3b0f767a9201"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.964742 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" event={"ID":"5da2e110-e194-4b21-a995-e7fcd1bb1825","Type":"ContainerStarted","Data":"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.965918 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.967782 4972 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-525bc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.967825 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.969865 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" event={"ID":"dde838bc-7c41-4da3-a00d-bb704e436faf","Type":"ContainerStarted","Data":"6788e52a2d1f4fde8c8e40b952f5901a77d5f4effd4e76d1e7d77aace5dbee44"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.975759 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" event={"ID":"63b91360-360a-46bf-ab58-6124f4210b84","Type":"ContainerStarted","Data":"74c43d8fa831d48fa707a25d4711ea4f0400cfdc6881e156976b6d5f541110fe"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.977699 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" event={"ID":"0dbb4a63-be87-47b1-8650-59c40ed5f7b5","Type":"ContainerStarted","Data":"9c0fed6158775e65627b158bdab548da2b431a7284c8dac657ab4f7484c411be"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.978659 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.978925 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.478898136 +0000 UTC m=+123.390905874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.978972 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:45 crc kubenswrapper[4972]: E0228 10:37:45.979532 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.479525105 +0000 UTC m=+123.391532843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.979543 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" event={"ID":"9bb2eeb1-2033-4895-af78-9cf70134182e","Type":"ContainerStarted","Data":"2fc6bf414b081bd3f5a6a3dcccfa72ac732a777b9a273e3fd494cdc8b6216375"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.979693 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.981682 4972 generic.go:334] "Generic (PLEG): container finished" podID="40d5e1c5-21a4-419a-9425-6773e3f33c29" containerID="97f0ddb30784ea22c0712c7d60125b3405ce3bd5c83cf0094ae59c22ad078b07" exitCode=0 Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.982501 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" event={"ID":"40d5e1c5-21a4-419a-9425-6773e3f33c29","Type":"ContainerDied","Data":"97f0ddb30784ea22c0712c7d60125b3405ce3bd5c83cf0094ae59c22ad078b07"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.988846 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" podStartSLOduration=53.988818411 podStartE2EDuration="53.988818411s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:45.986162942 +0000 UTC m=+122.898170680" watchObservedRunningTime="2026-02-28 10:37:45.988818411 +0000 UTC m=+122.900826149" Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.992884 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" event={"ID":"3c1531f8-6f83-4253-a88f-b2ec8144f50d","Type":"ContainerStarted","Data":"65123991af9e7235add2b133655727825faa73291994500c0c9df5f483bc2837"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.997074 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" event={"ID":"cc45436c-617d-4941-a3ab-952c1bfd4d4f","Type":"ContainerStarted","Data":"5611ba49fde34fc3f064f5254ff727998217c596b4db2a6b85625aa1cfadc591"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.998505 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" event={"ID":"06ae4169-c713-4ebc-88aa-dee750386952","Type":"ContainerStarted","Data":"2c5beb37d85b656b6252cc584bd1677dc76549e9ae5e3c30bb1ce2a1aeb03c23"} Feb 28 10:37:45 crc kubenswrapper[4972]: I0228 10:37:45.998808 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.001543 4972 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p5xl7 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.001601 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" podUID="06ae4169-c713-4ebc-88aa-dee750386952" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.005587 4972 generic.go:334] "Generic (PLEG): container finished" podID="9b27a0f8-30a8-4798-b7e7-612aa8313c2d" containerID="ad5eb64c05a5f4c41a428563d7f8b37acd16a10d3c20b09b71b64d8390234071" exitCode=0 Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.005695 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" event={"ID":"9b27a0f8-30a8-4798-b7e7-612aa8313c2d","Type":"ContainerDied","Data":"ad5eb64c05a5f4c41a428563d7f8b37acd16a10d3c20b09b71b64d8390234071"} Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.012641 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a562fc0b32ac8e60695ce507183c68f289a64627260995bce22977692975ada8"} Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.032989 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" event={"ID":"eeb9fc08-9a96-445c-9e73-dd9a2b796d17","Type":"ContainerStarted","Data":"ffe91705f500299d3426ab4074b88b74b1646aec61b9b72c6f1f2f1ac49f1f87"} Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.033948 4972 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-m6kfd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.034044 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.034639 4972 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9f545 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.037323 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Feb 28 10:37:46 crc kubenswrapper[4972]: W0228 10:37:46.040863 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-4e888ebc93093887352928521d5680c9f4e6fe91a6df359a3565b4225ba10c64 WatchSource:0}: Error finding container 4e888ebc93093887352928521d5680c9f4e6fe91a6df359a3565b4225ba10c64: Status 404 returned error can't find the container with id 4e888ebc93093887352928521d5680c9f4e6fe91a6df359a3565b4225ba10c64 Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.046562 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" podStartSLOduration=54.046540883 podStartE2EDuration="54.046540883s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:46.045017678 +0000 UTC m=+122.957025456" watchObservedRunningTime="2026-02-28 10:37:46.046540883 +0000 UTC m=+122.958548621" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.060083 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" podStartSLOduration=54.060065694 podStartE2EDuration="54.060065694s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:46.059798666 +0000 UTC m=+122.971806404" watchObservedRunningTime="2026-02-28 10:37:46.060065694 +0000 UTC m=+122.972073432" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.079983 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.080677 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.580654954 +0000 UTC m=+123.492662702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.105136 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpbj" podStartSLOduration=54.105085269 podStartE2EDuration="54.105085269s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:46.099662989 +0000 UTC m=+123.011670737" watchObservedRunningTime="2026-02-28 10:37:46.105085269 +0000 UTC m=+123.017093007" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.122540 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" podStartSLOduration=54.122516536 podStartE2EDuration="54.122516536s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:46.119706023 +0000 UTC m=+123.031713781" watchObservedRunningTime="2026-02-28 10:37:46.122516536 +0000 UTC m=+123.034524264" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.183125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.186629 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.686611828 +0000 UTC m=+123.598619566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.284636 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.284848 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.78481723 +0000 UTC m=+123.696824968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.285077 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.285514 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.785494781 +0000 UTC m=+123.697502519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.386206 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.386518 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.886448496 +0000 UTC m=+123.798456234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.386604 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.387334 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.887323022 +0000 UTC m=+123.799330760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.490326 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.990293557 +0000 UTC m=+123.902301295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.489511 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.492301 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.493145 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:46.99310477 +0000 UTC m=+123.905112508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.596190 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.596843 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.096821227 +0000 UTC m=+124.008828965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.699849 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.199830802 +0000 UTC m=+124.111838540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.700065 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.802097 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.802400 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.302384325 +0000 UTC m=+124.214392063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.809140 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.809199 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 28 10:37:46 crc kubenswrapper[4972]: I0228 10:37:46.903445 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:46 crc kubenswrapper[4972]: E0228 10:37:46.904042 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.404018579 +0000 UTC m=+124.316026307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.004380 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.004702 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.504686065 +0000 UTC m=+124.416693803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.042212 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a5b9e29f75f0b217b043348fa1e1439e8586c50861310e8114f04ee862f62c37"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.044644 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" event={"ID":"121139d5-7a49-4920-ac95-fc77bfe36b58","Type":"ContainerStarted","Data":"0eb9ccadecd229b94d4c2ad22d6d7cd23e3a68c6050c6b34e0c3e46a796ff267"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.046567 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lcv58" event={"ID":"b7f1c1c7-f8f6-418e-8526-e7946891f02e","Type":"ContainerStarted","Data":"318ede40b26e1ff17758f670e8280f49d3ed5faba206c7cd1441e56314217248"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.048993 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" event={"ID":"043d7cc4-5732-49b8-8150-bb897c5fa424","Type":"ContainerStarted","Data":"dd9e9b4a214cf42bd5d22c16b7cc44c02e99758df803149a7dbb403ce49b729d"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.050871 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" event={"ID":"fcd06083-d0dc-4523-b832-471ca8b45e82","Type":"ContainerStarted","Data":"600b0fcd008114f471becde7f895cd44c9e9693c6830e281dbc49f3c81b29cc6"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.052188 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4e888ebc93093887352928521d5680c9f4e6fe91a6df359a3565b4225ba10c64"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.054025 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" event={"ID":"cc45436c-617d-4941-a3ab-952c1bfd4d4f","Type":"ContainerStarted","Data":"c3f5423b220caf6d7574d4852e41ab1dfbeb14652583725b5d5708d51e1e23ef"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.055907 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a4c045b4dcdb75fea7f39e93a91d1fa73cfea449addc1af4841195768dbaed3a"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.055936 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d209ec30bb21a434e68b1279104dfb44fe7dc74772b3685b802d5a177c1e0988"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.059095 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" event={"ID":"c9173e02-eddc-40e3-86ab-0f0e8a8f50ac","Type":"ContainerStarted","Data":"d64b6f94bfff849e16ec4d78bbf20c3bfa2d6fc14aeba79e0fd98484458a5704"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.062335 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" event={"ID":"f7d82d58-d793-4201-b1c0-4ad0b6c28a20","Type":"ContainerStarted","Data":"d47fb3ee330ed9670d4e86e747b3b7115f88a022ae8c38d62b0b31ec5a994ecf"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.064505 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zxcst" event={"ID":"b4089e4d-bcd6-4524-bd70-b8eaf3768056","Type":"ContainerStarted","Data":"bf24558a1d4c285e0e523623a98edf9cfb57bbe73160dd108cac4232261c01fb"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.065948 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" event={"ID":"8066e21c-fa0c-497d-8bc8-a27d07dab5b4","Type":"ContainerStarted","Data":"5a6ff11997269a5c7dbae5275f5f58e0797ac891ef85246eeb8931ef1b129f40"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.068504 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" event={"ID":"c6085b35-69c1-40e1-a782-d9f7f5708181","Type":"ContainerStarted","Data":"0f719aa6febb6580320b32a1d0a40a080174dc136a181a32e0eb677bea92329b"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.071383 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" event={"ID":"ccfa71dd-d16e-482e-88f0-2b1a634f44b6","Type":"ContainerStarted","Data":"ed5c44aafeca17427f917ea26d165f7eef579a44d20ba4a068b24b1a55a1e5c5"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.074733 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" event={"ID":"79293b7f-3383-4376-90ca-1f6c6eff2dc4","Type":"ContainerStarted","Data":"04605804b4ee801f635f7e6f62a562d61ae2e10c079fe0406ce78db33c8d976d"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.078540 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" event={"ID":"1efe1619-53ab-4d7b-83d8-c772875cf078","Type":"ContainerStarted","Data":"0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.080699 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nkn8q" event={"ID":"699276a9-dfad-4ed8-9dcf-e197aa1d0a10","Type":"ContainerStarted","Data":"a58fd26694cdb5b932087ded7c07e03ae938d74d6c94c16eabbfd9624c71ef47"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.082248 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" event={"ID":"6bb9bb11-63e1-4af7-bef6-71f60cf33766","Type":"ContainerStarted","Data":"ae5841456b670a0999c2db45b14fdc782df95037bbb31e824850e65fe575a877"} Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.083924 4972 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p5xl7 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.083981 4972 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-9f545 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.084000 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" podUID="06ae4169-c713-4ebc-88aa-dee750386952" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.084052 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.084203 4972 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-525bc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.084234 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.094531 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7gn5m" podStartSLOduration=55.09451536 podStartE2EDuration="55.09451536s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.088635676 +0000 UTC m=+124.000643414" watchObservedRunningTime="2026-02-28 10:37:47.09451536 +0000 UTC m=+124.006523098" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.106215 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.106556 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.606542948 +0000 UTC m=+124.518550686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.109048 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xstmm" podStartSLOduration=55.109034151 podStartE2EDuration="55.109034151s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.10868256 +0000 UTC m=+124.020690298" watchObservedRunningTime="2026-02-28 10:37:47.109034151 +0000 UTC m=+124.021041889" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.130084 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x7r78" podStartSLOduration=55.130064425 podStartE2EDuration="55.130064425s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.128243821 +0000 UTC m=+124.040251559" watchObservedRunningTime="2026-02-28 10:37:47.130064425 +0000 UTC m=+124.042072163" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.150530 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" podStartSLOduration=55.150508331 podStartE2EDuration="55.150508331s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.14640575 +0000 UTC m=+124.058413488" watchObservedRunningTime="2026-02-28 10:37:47.150508331 +0000 UTC m=+124.062516069" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.166061 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bdb62" podStartSLOduration=55.166036892 podStartE2EDuration="55.166036892s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.163723034 +0000 UTC m=+124.075730772" watchObservedRunningTime="2026-02-28 10:37:47.166036892 +0000 UTC m=+124.078044630" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.195604 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-nfhbz" podStartSLOduration=7.195586618 podStartE2EDuration="7.195586618s" podCreationTimestamp="2026-02-28 10:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:47.193980161 +0000 UTC m=+124.105987899" watchObservedRunningTime="2026-02-28 10:37:47.195586618 +0000 UTC m=+124.107594356" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.207254 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.211045 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.711000506 +0000 UTC m=+124.623008244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.211272 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.225322 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.72528792 +0000 UTC m=+124.637295658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.313204 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.313668 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.813649031 +0000 UTC m=+124.725656769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.415786 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.416369 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:47.916347007 +0000 UTC m=+124.828354745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.517930 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.518133 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.018086636 +0000 UTC m=+124.930094374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.518383 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.518858 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.018847528 +0000 UTC m=+124.930855266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.619187 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.619533 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.119516555 +0000 UTC m=+125.031524293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.721158 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.721776 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.221752617 +0000 UTC m=+125.133760355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.808060 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.808117 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.821793 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.822403 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.822637 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.322609539 +0000 UTC m=+125.234617277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.822754 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.823116 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.323106093 +0000 UTC m=+125.235113831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:47 crc kubenswrapper[4972]: I0228 10:37:47.924787 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:47 crc kubenswrapper[4972]: E0228 10:37:47.925547 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.425520182 +0000 UTC m=+125.337527920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.026602 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.026952 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.52693989 +0000 UTC m=+125.438947628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.090923 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" event={"ID":"9b27a0f8-30a8-4798-b7e7-612aa8313c2d","Type":"ContainerStarted","Data":"33f264e4a9a28490e2ffc3010b4164f1bb8e06e0f1e5d9d60b69e72dddcb9e87"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.093115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e76c417200d0bad14ee1314611b3e1e0ebef910d93b9830dc4faa87559c04c44"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.095870 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" event={"ID":"ccfa71dd-d16e-482e-88f0-2b1a634f44b6","Type":"ContainerStarted","Data":"acc49ad9a78daad4924837465ae1aad7696723f724eba9ff252c56007785d948"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.099246 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" event={"ID":"40d5e1c5-21a4-419a-9425-6773e3f33c29","Type":"ContainerStarted","Data":"0ac5f6597bcc3158bec0875c7f8c1a94290d1ebc84c2a2ddaa0228318a7db574"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.099307 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" event={"ID":"40d5e1c5-21a4-419a-9425-6773e3f33c29","Type":"ContainerStarted","Data":"0ad2edd7708ba39e2ce68e4fe713be085d682767cf3a2a5e743fd475e5ab220f"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.103128 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" event={"ID":"6ba51b61-6394-4c9d-a5fb-389b2420f6e0","Type":"ContainerStarted","Data":"79a9f53414e7b1f2bc8301db2b7eebcd522ae52ce2f40fae2557f1ce7d0ef867"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.105924 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lcv58" event={"ID":"b7f1c1c7-f8f6-418e-8526-e7946891f02e","Type":"ContainerStarted","Data":"11a4fc8e8bc46b9d473077fde651b5b7c589562dec75b20b86d0d1c9df15addc"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.111668 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zxcst" event={"ID":"b4089e4d-bcd6-4524-bd70-b8eaf3768056","Type":"ContainerStarted","Data":"3c3a2b9c0d1f04d890c3f6c25820bbfc97dbf216f4a6b356a8d137bc0770ae58"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.111996 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.115046 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" event={"ID":"63b91360-360a-46bf-ab58-6124f4210b84","Type":"ContainerStarted","Data":"de058cbac09b8b636aa8cec947d3f4571a18a016bc2ad6f2aacff865e19b76f6"} Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.115940 4972 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-525bc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.115997 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.120168 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" podStartSLOduration=56.120121564 podStartE2EDuration="56.120121564s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.119195467 +0000 UTC m=+125.031203215" watchObservedRunningTime="2026-02-28 10:37:48.120121564 +0000 UTC m=+125.032129302" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.128283 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.128535 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.628496053 +0000 UTC m=+125.540503801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.128658 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.129138 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.629124122 +0000 UTC m=+125.541131860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.156882 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" podStartSLOduration=56.156861065 podStartE2EDuration="56.156861065s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.155973468 +0000 UTC m=+125.067981206" watchObservedRunningTime="2026-02-28 10:37:48.156861065 +0000 UTC m=+125.068868803" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.204919 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" podStartSLOduration=56.20489827 podStartE2EDuration="56.20489827s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.203933571 +0000 UTC m=+125.115941309" watchObservedRunningTime="2026-02-28 10:37:48.20489827 +0000 UTC m=+125.116906008" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.228291 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" podStartSLOduration=56.228268553 podStartE2EDuration="56.228268553s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.226012366 +0000 UTC m=+125.138020094" watchObservedRunningTime="2026-02-28 10:37:48.228268553 +0000 UTC m=+125.140276291" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.229788 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.229999 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.729965573 +0000 UTC m=+125.641973321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.230722 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.233016 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.732994983 +0000 UTC m=+125.645002721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.274528 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" podStartSLOduration=56.274503445 podStartE2EDuration="56.274503445s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.24939672 +0000 UTC m=+125.161404478" watchObservedRunningTime="2026-02-28 10:37:48.274503445 +0000 UTC m=+125.186511183" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.274660 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podStartSLOduration=8.274653589 podStartE2EDuration="8.274653589s" podCreationTimestamp="2026-02-28 10:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.270969539 +0000 UTC m=+125.182977277" watchObservedRunningTime="2026-02-28 10:37:48.274653589 +0000 UTC m=+125.186661327" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.288389 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xwmd" podStartSLOduration=56.288371205 podStartE2EDuration="56.288371205s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.286559902 +0000 UTC m=+125.198567630" watchObservedRunningTime="2026-02-28 10:37:48.288371205 +0000 UTC m=+125.200378943" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.320012 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gmfff" podStartSLOduration=56.319983984 podStartE2EDuration="56.319983984s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.311512792 +0000 UTC m=+125.223520520" watchObservedRunningTime="2026-02-28 10:37:48.319983984 +0000 UTC m=+125.231991722" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.335519 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.335709 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.835678859 +0000 UTC m=+125.747686597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.335830 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.336302 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.836293658 +0000 UTC m=+125.748301396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.359242 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nkn8q" podStartSLOduration=8.359215297 podStartE2EDuration="8.359215297s" podCreationTimestamp="2026-02-28 10:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.346113299 +0000 UTC m=+125.258121047" watchObservedRunningTime="2026-02-28 10:37:48.359215297 +0000 UTC m=+125.271223035" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.385155 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zxcst" podStartSLOduration=8.385129466 podStartE2EDuration="8.385129466s" podCreationTimestamp="2026-02-28 10:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.384865538 +0000 UTC m=+125.296873296" watchObservedRunningTime="2026-02-28 10:37:48.385129466 +0000 UTC m=+125.297137204" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.437430 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.437637 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.937604723 +0000 UTC m=+125.849612461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.437740 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.438069 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:48.938054776 +0000 UTC m=+125.850062514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.502997 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lcv58" podStartSLOduration=56.502977762 podStartE2EDuration="56.502977762s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.501806257 +0000 UTC m=+125.413813985" watchObservedRunningTime="2026-02-28 10:37:48.502977762 +0000 UTC m=+125.414985510" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.505084 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhfd6" podStartSLOduration=56.505070864 podStartE2EDuration="56.505070864s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.482404492 +0000 UTC m=+125.394412230" watchObservedRunningTime="2026-02-28 10:37:48.505070864 +0000 UTC m=+125.417078602" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.531547 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gzmqk" podStartSLOduration=56.531527218 podStartE2EDuration="56.531527218s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.530665953 +0000 UTC m=+125.442673691" watchObservedRunningTime="2026-02-28 10:37:48.531527218 +0000 UTC m=+125.443534956" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.539258 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.539381 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.039364401 +0000 UTC m=+125.951372139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.539639 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.540084 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.040064552 +0000 UTC m=+125.952072290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.605278 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bs4ks" podStartSLOduration=56.605258236 podStartE2EDuration="56.605258236s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.604596486 +0000 UTC m=+125.516604234" watchObservedRunningTime="2026-02-28 10:37:48.605258236 +0000 UTC m=+125.517265974" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.607415 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-44c45" podStartSLOduration=56.607400499 podStartE2EDuration="56.607400499s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.569772433 +0000 UTC m=+125.481780171" watchObservedRunningTime="2026-02-28 10:37:48.607400499 +0000 UTC m=+125.519408237" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.633902 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2ghv" podStartSLOduration=56.633885265 podStartE2EDuration="56.633885265s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.633204035 +0000 UTC m=+125.545211773" watchObservedRunningTime="2026-02-28 10:37:48.633885265 +0000 UTC m=+125.545893003" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.641106 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.641411 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.141373257 +0000 UTC m=+126.053380995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.641556 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.641918 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.141907853 +0000 UTC m=+126.053915601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.713900 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fs5p9" podStartSLOduration=56.713866587 podStartE2EDuration="56.713866587s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.71022447 +0000 UTC m=+125.622232208" watchObservedRunningTime="2026-02-28 10:37:48.713866587 +0000 UTC m=+125.625874325" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.715267 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vjxlj" podStartSLOduration=56.715260999 podStartE2EDuration="56.715260999s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:48.66843157 +0000 UTC m=+125.580439328" watchObservedRunningTime="2026-02-28 10:37:48.715260999 +0000 UTC m=+125.627268737" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.742424 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.742659 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.242620441 +0000 UTC m=+126.154628179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.742743 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.743181 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.243153767 +0000 UTC m=+126.155161505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.748940 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59646: no serving certificate available for the kubelet" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.818280 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:48 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:48 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:48 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.818427 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.844081 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.844238 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.344207694 +0000 UTC m=+126.256215432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.844298 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.844611 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.344603426 +0000 UTC m=+126.256611164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.847751 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59658: no serving certificate available for the kubelet" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.934124 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59674: no serving certificate available for the kubelet" Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.946019 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.946218 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.446183279 +0000 UTC m=+126.358191017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.946404 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:48 crc kubenswrapper[4972]: E0228 10:37:48.946836 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.446825319 +0000 UTC m=+126.358833067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:48 crc kubenswrapper[4972]: I0228 10:37:48.990637 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59676: no serving certificate available for the kubelet" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.047360 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.047585 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.547555426 +0000 UTC m=+126.459563174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.047702 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.048004 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.54799665 +0000 UTC m=+126.460004388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.089045 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59686: no serving certificate available for the kubelet" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.149050 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.149199 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.649178301 +0000 UTC m=+126.561186039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.149406 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.149710 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.649702417 +0000 UTC m=+126.561710155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.167100 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59688: no serving certificate available for the kubelet" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.252686 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.254735 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.754713201 +0000 UTC m=+126.666720939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.257696 4972 ???:1] "http: TLS handshake error from 192.168.126.11:59698: no serving certificate available for the kubelet" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.356087 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.356734 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.856708138 +0000 UTC m=+126.768715866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.457038 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.457271 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.95722701 +0000 UTC m=+126.869234748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.457359 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.457766 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:49.957745945 +0000 UTC m=+126.869753683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.558678 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.558864 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.058837494 +0000 UTC m=+126.970845232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.558935 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.559218 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.059210484 +0000 UTC m=+126.971218222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.622186 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58676: no serving certificate available for the kubelet" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.659816 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.660152 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.160101597 +0000 UTC m=+127.072109345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.660235 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.660599 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.160580071 +0000 UTC m=+127.072587809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.761121 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.761354 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.26131973 +0000 UTC m=+127.173327478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.761406 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.761824 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.261814224 +0000 UTC m=+127.173821962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.810997 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:49 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:49 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:49 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.811074 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.862618 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.862792 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.362758839 +0000 UTC m=+127.274766577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.862826 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.863418 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.363405189 +0000 UTC m=+127.275412927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.964083 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.964261 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.46424113 +0000 UTC m=+127.376248868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:49 crc kubenswrapper[4972]: I0228 10:37:49.964345 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:49 crc kubenswrapper[4972]: E0228 10:37:49.964660 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.464653962 +0000 UTC m=+127.376661700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.065879 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.066139 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.566098621 +0000 UTC m=+127.478106359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.066933 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.067384 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.56737435 +0000 UTC m=+127.479382088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.136835 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" event={"ID":"7b440217-3770-4ec9-b252-868fe01353fa","Type":"ContainerStarted","Data":"4b502405b2ec7ea057ac71aed9131a5d5750b0aacb137fc988b9e1366da0572d"} Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.168738 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.169159 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.669136378 +0000 UTC m=+127.581144106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.271239 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.771223256 +0000 UTC m=+127.683230994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.270814 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.373166 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.373328 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.873302975 +0000 UTC m=+127.785310713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.373420 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.373720 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.873712887 +0000 UTC m=+127.785720615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.386549 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58690: no serving certificate available for the kubelet" Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.474790 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.475015 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.97498098 +0000 UTC m=+127.886988718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.475200 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.475809 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:50.975801155 +0000 UTC m=+127.887808893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.576802 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.577047 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.076992036 +0000 UTC m=+127.988999794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.577728 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.578178 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.078163672 +0000 UTC m=+127.990171410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.679007 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.679204 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.179177028 +0000 UTC m=+128.091184766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.679270 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.679611 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.1796024 +0000 UTC m=+128.091610138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.780317 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.780576 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.280525154 +0000 UTC m=+128.192532892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.780741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.781263 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.281253906 +0000 UTC m=+128.193261644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.810904 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:50 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:50 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:50 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.810986 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.881907 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.882077 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.382049226 +0000 UTC m=+128.294056964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.882204 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.882518 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.382503859 +0000 UTC m=+128.294511587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.982771 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.983039 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.48298457 +0000 UTC m=+128.394992318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:50 crc kubenswrapper[4972]: I0228 10:37:50.983138 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:50 crc kubenswrapper[4972]: E0228 10:37:50.983491 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.483475095 +0000 UTC m=+128.395482823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.084556 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.084768 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.584742069 +0000 UTC m=+128.496749807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.084911 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.085242 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.585228773 +0000 UTC m=+128.497236511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.122659 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.123292 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.128125 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.128331 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.153424 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.170004 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.170598 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" containerID="cri-o://bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a" gracePeriod=30 Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.181148 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.181366 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" containerID="cri-o://69d364177737fc390b663d73b69d7071a89b5009461a712a071667ca821c8333" gracePeriod=30 Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.185881 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.186220 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.686205039 +0000 UTC m=+128.598212777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.191290 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.193234 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.287243 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.287350 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.287430 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.287824 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.787809053 +0000 UTC m=+128.699816791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.389003 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.389220 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.88918634 +0000 UTC m=+128.801194088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.389555 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.389598 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.389636 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.390020 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.390252 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.890240351 +0000 UTC m=+128.802248089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.442960 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.491401 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.491624 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.991592168 +0000 UTC m=+128.903599906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.491830 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.492190 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:51.992180165 +0000 UTC m=+128.904187903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.594024 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.594456 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.094435289 +0000 UTC m=+129.006443027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.696679 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.697100 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.197084593 +0000 UTC m=+129.109092331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.713061 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bd989" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.734864 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58698: no serving certificate available for the kubelet" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.740962 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.763361 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.797841 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.798608 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.298586454 +0000 UTC m=+129.210594192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.841298 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:51 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:51 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:51 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.841374 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.899320 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert\") pod \"5da2e110-e194-4b21-a995-e7fcd1bb1825\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.899397 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69btw\" (UniqueName: \"kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw\") pod \"5da2e110-e194-4b21-a995-e7fcd1bb1825\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.899498 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config\") pod \"5da2e110-e194-4b21-a995-e7fcd1bb1825\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.899554 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca\") pod \"5da2e110-e194-4b21-a995-e7fcd1bb1825\" (UID: \"5da2e110-e194-4b21-a995-e7fcd1bb1825\") " Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.899937 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.901451 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config" (OuterVolumeSpecName: "config") pod "5da2e110-e194-4b21-a995-e7fcd1bb1825" (UID: "5da2e110-e194-4b21-a995-e7fcd1bb1825"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.901737 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca" (OuterVolumeSpecName: "client-ca") pod "5da2e110-e194-4b21-a995-e7fcd1bb1825" (UID: "5da2e110-e194-4b21-a995-e7fcd1bb1825"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:51 crc kubenswrapper[4972]: E0228 10:37:51.902237 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.402224179 +0000 UTC m=+129.314231917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.906961 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw" (OuterVolumeSpecName: "kube-api-access-69btw") pod "5da2e110-e194-4b21-a995-e7fcd1bb1825" (UID: "5da2e110-e194-4b21-a995-e7fcd1bb1825"). InnerVolumeSpecName "kube-api-access-69btw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.917558 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5da2e110-e194-4b21-a995-e7fcd1bb1825" (UID: "5da2e110-e194-4b21-a995-e7fcd1bb1825"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.947503 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.947584 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.947974 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:37:51 crc kubenswrapper[4972]: I0228 10:37:51.948019 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.000603 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.000936 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69btw\" (UniqueName: \"kubernetes.io/projected/5da2e110-e194-4b21-a995-e7fcd1bb1825-kube-api-access-69btw\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.000950 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.000958 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5da2e110-e194-4b21-a995-e7fcd1bb1825-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.000969 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5da2e110-e194-4b21-a995-e7fcd1bb1825-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.001026 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.500995749 +0000 UTC m=+129.413003487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.122207 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.123036 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.623012968 +0000 UTC m=+129.535020706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.181529 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.192870 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.192918 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.192957 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.218796 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5d29s" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.225828 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.228162 4972 generic.go:334] "Generic (PLEG): container finished" podID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerID="69d364177737fc390b663d73b69d7071a89b5009461a712a071667ca821c8333" exitCode=0 Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.228325 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" event={"ID":"0d27c7ba-b3c5-492a-9a57-53a179cc40d9","Type":"ContainerDied","Data":"69d364177737fc390b663d73b69d7071a89b5009461a712a071667ca821c8333"} Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.229432 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.729411715 +0000 UTC m=+129.641419453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.235474 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.235914 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.735892978 +0000 UTC m=+129.647900716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.237882 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.246344 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.246597 4972 generic.go:334] "Generic (PLEG): container finished" podID="dde838bc-7c41-4da3-a00d-bb704e436faf" containerID="6788e52a2d1f4fde8c8e40b952f5901a77d5f4effd4e76d1e7d77aace5dbee44" exitCode=0 Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.246713 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" event={"ID":"dde838bc-7c41-4da3-a00d-bb704e436faf","Type":"ContainerDied","Data":"6788e52a2d1f4fde8c8e40b952f5901a77d5f4effd4e76d1e7d77aace5dbee44"} Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.247085 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.251141 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.251509 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.251529 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.251687 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerName="route-controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.258211 4972 patch_prober.go:28] interesting pod/console-f9d7485db-czzdx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.258261 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-czzdx" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.260896 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.263334 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.265309 4972 generic.go:334] "Generic (PLEG): container finished" podID="5da2e110-e194-4b21-a995-e7fcd1bb1825" containerID="bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a" exitCode=0 Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.265364 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" event={"ID":"5da2e110-e194-4b21-a995-e7fcd1bb1825","Type":"ContainerDied","Data":"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a"} Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.265407 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" event={"ID":"5da2e110-e194-4b21-a995-e7fcd1bb1825","Type":"ContainerDied","Data":"e82a11a53798b6c922b808cb6ed6033a56bf3755640ac5d0e695a15245956f54"} Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.265432 4972 scope.go:117] "RemoveContainer" containerID="bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.265856 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.292122 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.331590 4972 scope.go:117] "RemoveContainer" containerID="bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.338947 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a\": container with ID starting with bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a not found: ID does not exist" containerID="bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.339009 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a"} err="failed to get container status \"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a\": rpc error: code = NotFound desc = could not find container \"bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a\": container with ID starting with bb006bd9d6f0da11fc3ce9729a7d66bfdc765d98da3e31a09033da06156d9e9a not found: ID does not exist" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.339601 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.340638 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.840605073 +0000 UTC m=+129.752612811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.360946 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.370509 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-525bc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.416565 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.425530 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.425827 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.425848 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.425991 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.426887 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.429947 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.442571 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgw7r\" (UniqueName: \"kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.442693 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.442728 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.442780 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.444045 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:52.944027452 +0000 UTC m=+129.856035190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.445254 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.538861 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.539593 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.540981 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.541387 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.541508 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.541624 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.542630 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.543427 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.543975 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert\") pod \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544011 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ktd2\" (UniqueName: \"kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2\") pod \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544038 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config\") pod \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544085 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544106 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544153 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544179 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca\") pod \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544222 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles\") pod \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\" (UID: \"0d27c7ba-b3c5-492a-9a57-53a179cc40d9\") " Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.544430 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.044403549 +0000 UTC m=+129.956411287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544488 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544550 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgw7r\" (UniqueName: \"kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544617 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544658 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544763 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.544805 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgpd7\" (UniqueName: \"kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.545321 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config" (OuterVolumeSpecName: "config") pod "0d27c7ba-b3c5-492a-9a57-53a179cc40d9" (UID: "0d27c7ba-b3c5-492a-9a57-53a179cc40d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.545333 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.045325886 +0000 UTC m=+129.957333624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.545747 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0d27c7ba-b3c5-492a-9a57-53a179cc40d9" (UID: "0d27c7ba-b3c5-492a-9a57-53a179cc40d9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.546034 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.546105 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.549165 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca" (OuterVolumeSpecName: "client-ca") pod "0d27c7ba-b3c5-492a-9a57-53a179cc40d9" (UID: "0d27c7ba-b3c5-492a-9a57-53a179cc40d9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.556068 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2" (OuterVolumeSpecName: "kube-api-access-9ktd2") pod "0d27c7ba-b3c5-492a-9a57-53a179cc40d9" (UID: "0d27c7ba-b3c5-492a-9a57-53a179cc40d9"). InnerVolumeSpecName "kube-api-access-9ktd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.559793 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.561822 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.563774 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0d27c7ba-b3c5-492a-9a57-53a179cc40d9" (UID: "0d27c7ba-b3c5-492a-9a57-53a179cc40d9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.592836 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgw7r\" (UniqueName: \"kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r\") pod \"certified-operators-wzx7z\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.602285 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.604186 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.611950 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.612573 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.635576 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.643775 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650547 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650786 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr65z\" (UniqueName: \"kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650812 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650888 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5wpb\" (UniqueName: \"kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650927 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.650995 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgpd7\" (UniqueName: \"kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651025 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651040 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651163 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651197 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651215 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651229 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651320 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651387 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651448 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ktd2\" (UniqueName: \"kubernetes.io/projected/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-kube-api-access-9ktd2\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651493 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651503 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.651513 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d27c7ba-b3c5-492a-9a57-53a179cc40d9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.651692 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.15163732 +0000 UTC m=+130.063645058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.652172 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.653095 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.656537 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.659008 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.661438 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.669345 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgpd7\" (UniqueName: \"kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7\") pod \"community-operators-6jglk\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.753783 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755346 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755425 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755472 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755498 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755548 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755585 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755626 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755651 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr65z\" (UniqueName: \"kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755678 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755705 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5wpb\" (UniqueName: \"kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755736 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755768 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85kkr\" (UniqueName: \"kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755797 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755826 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.755875 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.758770 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.758929 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.258893142 +0000 UTC m=+130.170901070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.760077 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.761151 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.761200 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.761724 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.762068 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.762073 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.767566 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.767911 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.783117 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5wpb\" (UniqueName: \"kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb\") pod \"controller-manager-6487d4cd87-6gz5v\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.785181 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr65z\" (UniqueName: \"kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z\") pod \"route-controller-manager-74fbb5c645-9rjwr\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.806016 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.811809 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.815127 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:52 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:52 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:52 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.815201 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.824320 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.834948 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.847561 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p5xl7" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.856818 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.861172 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.869806 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870013 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870114 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgtk4\" (UniqueName: \"kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870145 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870182 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870237 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870285 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85kkr\" (UniqueName: \"kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870315 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.870416 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.870905 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.370884944 +0000 UTC m=+130.282892682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.871706 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.872304 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.873737 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.878851 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.899965 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85kkr\" (UniqueName: \"kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr\") pod \"certified-operators-dqghp\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.902199 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.912335 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.912693 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.926076 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.938547 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=0.93852959 podStartE2EDuration="938.52959ms" podCreationTimestamp="2026-02-28 10:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:52.906849231 +0000 UTC m=+129.818856979" watchObservedRunningTime="2026-02-28 10:37:52.93852959 +0000 UTC m=+129.850537328" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.953872 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.962356 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.974574 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.974679 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.974718 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgtk4\" (UniqueName: \"kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.974747 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.978035 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: E0228 10:37:52.978838 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.478822576 +0000 UTC m=+130.390830314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.979430 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:52 crc kubenswrapper[4972]: I0228 10:37:52.985960 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:37:52 crc kubenswrapper[4972]: W0228 10:37:52.988198 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod285bfdff_92e3_4ec5_bbd8_745860eef1fa.slice/crio-e8159e1c553323ded7f90e7db8a6d395cadfddbd50c736eed4227a737f84af38 WatchSource:0}: Error finding container e8159e1c553323ded7f90e7db8a6d395cadfddbd50c736eed4227a737f84af38: Status 404 returned error can't find the container with id e8159e1c553323ded7f90e7db8a6d395cadfddbd50c736eed4227a737f84af38 Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.000694 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgtk4\" (UniqueName: \"kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4\") pod \"community-operators-85647\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.075691 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.076407 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.576387481 +0000 UTC m=+130.488395219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.098237 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:37:53 crc kubenswrapper[4972]: W0228 10:37:53.119223 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66ead19f_03ec_41d1_82b4_9697bbb1e6a6.slice/crio-5227ff53a47d49168ed6a2712f2e0360ae1028ceabcfadf2bdfee23db3dcab65 WatchSource:0}: Error finding container 5227ff53a47d49168ed6a2712f2e0360ae1028ceabcfadf2bdfee23db3dcab65: Status 404 returned error can't find the container with id 5227ff53a47d49168ed6a2712f2e0360ae1028ceabcfadf2bdfee23db3dcab65 Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.124161 4972 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-m6kfd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: i/o timeout" start-of-body= Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.124744 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: i/o timeout" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.177663 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.178031 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.678017515 +0000 UTC m=+130.590025253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.183984 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85647" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.285876 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.286000 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.785978837 +0000 UTC m=+130.697986565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.286283 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.286585 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.786578136 +0000 UTC m=+130.698585864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.307710 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" event={"ID":"7b440217-3770-4ec9-b252-868fe01353fa","Type":"ContainerStarted","Data":"a405196097244cd6be0f2d118b338a3f8d41cb77f102069f6b1d508ccf84eb7c"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.310588 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerStarted","Data":"e8159e1c553323ded7f90e7db8a6d395cadfddbd50c736eed4227a737f84af38"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.312491 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.313088 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m6kfd" event={"ID":"0d27c7ba-b3c5-492a-9a57-53a179cc40d9","Type":"ContainerDied","Data":"3ded79de9651c6d1cebe41040b6cf621b6e1990131701708a1ad25a7627603eb"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.313120 4972 scope.go:117] "RemoveContainer" containerID="69d364177737fc390b663d73b69d7071a89b5009461a712a071667ca821c8333" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.315079 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerStarted","Data":"5227ff53a47d49168ed6a2712f2e0360ae1028ceabcfadf2bdfee23db3dcab65"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.329322 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.340342 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"df72f6a2-9aad-485e-8265-319ce6cb4315","Type":"ContainerStarted","Data":"bbb21f5a8b0d3c0bda1975d3cf66387a163ce2a2cf62cdf7c32e94c1393d47d9"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.340398 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"df72f6a2-9aad-485e-8265-319ce6cb4315","Type":"ContainerStarted","Data":"32c4692869ff4a22603465e3ebf2e81d242799507bce828c1018cd29f8544d39"} Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.358337 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6s2p6" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.388481 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.388892 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.888851469 +0000 UTC m=+130.800859207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.389078 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.393084 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.893057134 +0000 UTC m=+130.805064872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.393589 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.393562399 podStartE2EDuration="2.393562399s" podCreationTimestamp="2026-02-28 10:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:53.373887306 +0000 UTC m=+130.285895054" watchObservedRunningTime="2026-02-28 10:37:53.393562399 +0000 UTC m=+130.305570137" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.411176 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.478354 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.492826 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m6kfd"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.495846 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.496164 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.996115971 +0000 UTC m=+130.908123709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.496486 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.497035 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:53.997023258 +0000 UTC m=+130.909030996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.511862 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.515212 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.534293 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.544212 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wntf7" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.565641 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kcr2z" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.583069 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.599136 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.602242 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.102219588 +0000 UTC m=+131.014227326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.701566 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.703093 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.20306352 +0000 UTC m=+131.115071438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.802757 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.803190 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.303148529 +0000 UTC m=+131.215156267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.818172 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d27c7ba-b3c5-492a-9a57-53a179cc40d9" path="/var/lib/kubelet/pods/0d27c7ba-b3c5-492a-9a57-53a179cc40d9/volumes" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.819376 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:53 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:53 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:53 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.819439 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.819659 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da2e110-e194-4b21-a995-e7fcd1bb1825" path="/var/lib/kubelet/pods/5da2e110-e194-4b21-a995-e7fcd1bb1825/volumes" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.824014 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.825447 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.834587 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:37:53 crc kubenswrapper[4972]: W0228 10:37:53.864982 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cea766f_d665_474b_a7fa_5e193ce3c1fd.slice/crio-19fa698ae570b087cea9567de413259c2a108e2dc164e642a7701e366fca5b2f WatchSource:0}: Error finding container 19fa698ae570b087cea9567de413259c2a108e2dc164e642a7701e366fca5b2f: Status 404 returned error can't find the container with id 19fa698ae570b087cea9567de413259c2a108e2dc164e642a7701e366fca5b2f Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.909992 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:53 crc kubenswrapper[4972]: E0228 10:37:53.910449 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.410431781 +0000 UTC m=+131.322439519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.921117 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:53 crc kubenswrapper[4972]: I0228 10:37:53.982139 4972 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.017142 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.017345 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume\") pod \"dde838bc-7c41-4da3-a00d-bb704e436faf\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.017395 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume\") pod \"dde838bc-7c41-4da3-a00d-bb704e436faf\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.017446 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgbzv\" (UniqueName: \"kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv\") pod \"dde838bc-7c41-4da3-a00d-bb704e436faf\" (UID: \"dde838bc-7c41-4da3-a00d-bb704e436faf\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.017511 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.517480978 +0000 UTC m=+131.429488716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.017737 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.018887 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume" (OuterVolumeSpecName: "config-volume") pod "dde838bc-7c41-4da3-a00d-bb704e436faf" (UID: "dde838bc-7c41-4da3-a00d-bb704e436faf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.020369 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.520347933 +0000 UTC m=+131.432355671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.027230 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv" (OuterVolumeSpecName: "kube-api-access-mgbzv") pod "dde838bc-7c41-4da3-a00d-bb704e436faf" (UID: "dde838bc-7c41-4da3-a00d-bb704e436faf"). InnerVolumeSpecName "kube-api-access-mgbzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.038236 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dde838bc-7c41-4da3-a00d-bb704e436faf" (UID: "dde838bc-7c41-4da3-a00d-bb704e436faf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.118859 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.119223 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dde838bc-7c41-4da3-a00d-bb704e436faf-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.119245 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgbzv\" (UniqueName: \"kubernetes.io/projected/dde838bc-7c41-4da3-a00d-bb704e436faf-kube-api-access-mgbzv\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.119254 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dde838bc-7c41-4da3-a00d-bb704e436faf-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.119320 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.619302998 +0000 UTC m=+131.531310736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.162614 4972 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2xkbh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]log ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]etcd ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/generic-apiserver-start-informers ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/max-in-flight-filter ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 28 10:37:54 crc kubenswrapper[4972]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 28 10:37:54 crc kubenswrapper[4972]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/project.openshift.io-projectcache ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/openshift.io-startinformers ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 28 10:37:54 crc kubenswrapper[4972]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 28 10:37:54 crc kubenswrapper[4972]: livez check failed Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.162745 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" podUID="40d5e1c5-21a4-419a-9425-6773e3f33c29" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.220883 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.221304 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.721284253 +0000 UTC m=+131.633291991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.321969 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.322341 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.822303799 +0000 UTC m=+131.734311537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.322790 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.323112 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.823101484 +0000 UTC m=+131.735109222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.335028 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58712: no serving certificate available for the kubelet" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.345355 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" event={"ID":"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d","Type":"ContainerStarted","Data":"5fe16a52af070fda0971e1a4044dc0bd9ec3f3043226fbee05b02e279b221d78"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.347179 4972 generic.go:334] "Generic (PLEG): container finished" podID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerID="439742b31ac32f4e1d2846a1d3bcaeb86f15fb8a51aaf6af18a7ba949f5cd3c2" exitCode=0 Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.347263 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerDied","Data":"439742b31ac32f4e1d2846a1d3bcaeb86f15fb8a51aaf6af18a7ba949f5cd3c2"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.352106 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" event={"ID":"7b440217-3770-4ec9-b252-868fe01353fa","Type":"ContainerStarted","Data":"607db43e2958efb45a538287ec0f4ac99494d15b1ad224f94aab06beeb2d8c77"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.354377 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerStarted","Data":"325083a46718fc482f039ad27d5c92a2dfad1fce3cdb5dac23930b15949b1b98"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.356476 4972 generic.go:334] "Generic (PLEG): container finished" podID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerID="b038f7d2f38eb1192a0087044f7253e8d0d7e66580b84ee41835d76c6c071ecd" exitCode=0 Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.356602 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerDied","Data":"b038f7d2f38eb1192a0087044f7253e8d0d7e66580b84ee41835d76c6c071ecd"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.359875 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" event={"ID":"dde838bc-7c41-4da3-a00d-bb704e436faf","Type":"ContainerDied","Data":"152ec1bca1ca80037a09b0f50cbf0de64cbd2e68fdbb09c2d9fbb37ec0e88c37"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.359915 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="152ec1bca1ca80037a09b0f50cbf0de64cbd2e68fdbb09c2d9fbb37ec0e88c37" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.359978 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.367567 4972 generic.go:334] "Generic (PLEG): container finished" podID="df72f6a2-9aad-485e-8265-319ce6cb4315" containerID="bbb21f5a8b0d3c0bda1975d3cf66387a163ce2a2cf62cdf7c32e94c1393d47d9" exitCode=0 Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.367751 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"df72f6a2-9aad-485e-8265-319ce6cb4315","Type":"ContainerDied","Data":"bbb21f5a8b0d3c0bda1975d3cf66387a163ce2a2cf62cdf7c32e94c1393d47d9"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.376542 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerStarted","Data":"19fa698ae570b087cea9567de413259c2a108e2dc164e642a7701e366fca5b2f"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.378705 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" event={"ID":"ba81daec-9cd5-4f06-99e5-9f01366cf4bf","Type":"ContainerStarted","Data":"e7c1502771c6792973cc250a67981cd8e533e6c5991b0b14a198aca0605789eb"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.381383 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"84685205-887c-4e3b-8c9a-6951c278e2d6","Type":"ContainerStarted","Data":"9f57d799b1ca6a4657f173f515ddb48684309a72c582c93649eb6d37fb3d7398"} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.419016 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.419421 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.419782 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde838bc-7c41-4da3-a00d-bb704e436faf" containerName="collect-profiles" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.419799 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde838bc-7c41-4da3-a00d-bb704e436faf" containerName="collect-profiles" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.419924 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde838bc-7c41-4da3-a00d-bb704e436faf" containerName="collect-profiles" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.421025 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.424411 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.425250 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.426924 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:54.926902243 +0000 UTC m=+131.838909981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.455791 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.543842 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.543969 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzjp6\" (UniqueName: \"kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.544010 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.544056 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.546210 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.0461657 +0000 UTC m=+131.958173448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.646815 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.647099 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.147056183 +0000 UTC m=+132.059063921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.649889 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.650412 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.650634 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzjp6\" (UniqueName: \"kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.651089 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.651439 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.651515 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.651691 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.15167839 +0000 UTC m=+132.063686338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.684933 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzjp6\" (UniqueName: \"kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6\") pod \"redhat-marketplace-2j69x\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.752647 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.752841 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.25280179 +0000 UTC m=+132.164809528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.752934 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.753393 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.253375717 +0000 UTC m=+132.165383455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.810193 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:54 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:54 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:54 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.810273 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.810941 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.814007 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.815756 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.832359 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.853843 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.854012 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.353985622 +0000 UTC m=+132.265993360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.854262 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9dq9\" (UniqueName: \"kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.854372 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.854413 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.854443 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.854907 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.354897618 +0000 UTC m=+132.266905356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f7g8b" (UID: "8bc8f405-1a09-41db-9582-091d232da716") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.949179 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-5plg9"] Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.951267 4972 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-28T10:37:53.98217094Z","Handler":null,"Name":""} Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.955943 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.956130 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9dq9\" (UniqueName: \"kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.956203 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.956232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: E0228 10:37:54.956409 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-28 10:37:55.456351409 +0000 UTC m=+132.368359287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.956736 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.956886 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.965432 4972 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.965497 4972 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 28 10:37:54 crc kubenswrapper[4972]: I0228 10:37:54.974908 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9dq9\" (UniqueName: \"kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9\") pod \"redhat-marketplace-bccmr\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.062522 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.079356 4972 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.079400 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.103645 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.130979 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f7g8b\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.132323 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.140240 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:37:55 crc kubenswrapper[4972]: W0228 10:37:55.142778 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad7908e4_f813_4c17_a672_a9ff32bdfd8c.slice/crio-8ddce064c8404db9141d14e0c2464f891d20eaf08b073a0552568f8d9b323c71 WatchSource:0}: Error finding container 8ddce064c8404db9141d14e0c2464f891d20eaf08b073a0552568f8d9b323c71: Status 404 returned error can't find the container with id 8ddce064c8404db9141d14e0c2464f891d20eaf08b073a0552568f8d9b323c71 Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.164086 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.177740 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.224554 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.233322 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.394665 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" event={"ID":"ba81daec-9cd5-4f06-99e5-9f01366cf4bf","Type":"ContainerStarted","Data":"3592e67bc54524c45f6bdd3781ff8a3f244fb851f74357ad1d551a786e45b122"} Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.395848 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerStarted","Data":"8ddce064c8404db9141d14e0c2464f891d20eaf08b073a0552568f8d9b323c71"} Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.397185 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" gracePeriod=30 Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.397928 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"84685205-887c-4e3b-8c9a-6951c278e2d6","Type":"ContainerStarted","Data":"3040030d68b6a6a9d7490bf4ceddd86bb16cd677ea8bb22045902fd797c83c83"} Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.404106 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.420820 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.422220 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.428344 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.478952 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.496373 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.581397 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.581446 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlnmm\" (UniqueName: \"kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.581497 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.682833 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.682903 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlnmm\" (UniqueName: \"kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.682943 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.683490 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.683586 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.703288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlnmm\" (UniqueName: \"kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm\") pod \"redhat-operators-f45gq\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.763690 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.775196 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.804415 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.819108 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:55 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:55 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:55 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.819171 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.820925 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:37:55 crc kubenswrapper[4972]: E0228 10:37:55.823312 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df72f6a2-9aad-485e-8265-319ce6cb4315" containerName="pruner" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.823332 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="df72f6a2-9aad-485e-8265-319ce6cb4315" containerName="pruner" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.823525 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="df72f6a2-9aad-485e-8265-319ce6cb4315" containerName="pruner" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.827081 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.837007 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.894353 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access\") pod \"df72f6a2-9aad-485e-8265-319ce6cb4315\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.894439 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir\") pod \"df72f6a2-9aad-485e-8265-319ce6cb4315\" (UID: \"df72f6a2-9aad-485e-8265-319ce6cb4315\") " Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.894842 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l92mf\" (UniqueName: \"kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.894932 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.894974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.896432 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "df72f6a2-9aad-485e-8265-319ce6cb4315" (UID: "df72f6a2-9aad-485e-8265-319ce6cb4315"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.902393 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "df72f6a2-9aad-485e-8265-319ce6cb4315" (UID: "df72f6a2-9aad-485e-8265-319ce6cb4315"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.996367 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.996504 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l92mf\" (UniqueName: \"kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.996552 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.996587 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df72f6a2-9aad-485e-8265-319ce6cb4315-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.996598 4972 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/df72f6a2-9aad-485e-8265-319ce6cb4315-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.997138 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:55 crc kubenswrapper[4972]: I0228 10:37:55.997415 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.017074 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l92mf\" (UniqueName: \"kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf\") pod \"redhat-operators-w8pgq\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.023416 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:37:56 crc kubenswrapper[4972]: W0228 10:37:56.039673 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf90cea3_090d_4f18_88e1_c6de1a5fc469.slice/crio-799df91914a3dbe4e458d7c17041ff5ba28f9ca03f5299ff858b3cd47c2a5380 WatchSource:0}: Error finding container 799df91914a3dbe4e458d7c17041ff5ba28f9ca03f5299ff858b3cd47c2a5380: Status 404 returned error can't find the container with id 799df91914a3dbe4e458d7c17041ff5ba28f9ca03f5299ff858b3cd47c2a5380 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.209363 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.416366 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerID="d084c942847d9f0e5ebd73330e65697132aa4cc6c742ba91ecf668de5c4a3397" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.416776 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerDied","Data":"d084c942847d9f0e5ebd73330e65697132aa4cc6c742ba91ecf668de5c4a3397"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.438771 4972 generic.go:334] "Generic (PLEG): container finished" podID="84685205-887c-4e3b-8c9a-6951c278e2d6" containerID="3040030d68b6a6a9d7490bf4ceddd86bb16cd677ea8bb22045902fd797c83c83" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.438829 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"84685205-887c-4e3b-8c9a-6951c278e2d6","Type":"ContainerDied","Data":"3040030d68b6a6a9d7490bf4ceddd86bb16cd677ea8bb22045902fd797c83c83"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.456663 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerID="7e9b621b53bf5eb85267db6b753363f14407c592efb3f1b6505fe05889c82280" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.456830 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerDied","Data":"7e9b621b53bf5eb85267db6b753363f14407c592efb3f1b6505fe05889c82280"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.456881 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerStarted","Data":"dd9f8b89d6baf1bc65633b3e98f88fd92b66ef355bc37c2fa41ef829199cbb03"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.482019 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"df72f6a2-9aad-485e-8265-319ce6cb4315","Type":"ContainerDied","Data":"32c4692869ff4a22603465e3ebf2e81d242799507bce828c1018cd29f8544d39"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.482066 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32c4692869ff4a22603465e3ebf2e81d242799507bce828c1018cd29f8544d39" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.482141 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.505655 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerID="571c5d16bcd3380361cd3f5eaf9df084a384ecad6c6df862c61b54e19b0b41a2" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.505788 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerDied","Data":"571c5d16bcd3380361cd3f5eaf9df084a384ecad6c6df862c61b54e19b0b41a2"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.509578 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" event={"ID":"8bc8f405-1a09-41db-9582-091d232da716","Type":"ContainerStarted","Data":"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.509603 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" event={"ID":"8bc8f405-1a09-41db-9582-091d232da716","Type":"ContainerStarted","Data":"19adb258528068a3b37686fc0be619fbf16b396c7a729303015a0d00f064b4b9"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.509736 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.514025 4972 generic.go:334] "Generic (PLEG): container finished" podID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerID="11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.514230 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerDied","Data":"11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.544054 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" event={"ID":"7b440217-3770-4ec9-b252-868fe01353fa","Type":"ContainerStarted","Data":"146a01a35e16ba196dbd584f00fcd3b8423a6f5efc847585ea7e3bb472eb48c9"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.568516 4972 generic.go:334] "Generic (PLEG): container finished" podID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerID="6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665" exitCode=0 Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.568742 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerDied","Data":"6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.568791 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerStarted","Data":"799df91914a3dbe4e458d7c17041ff5ba28f9ca03f5299ff858b3cd47c2a5380"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.574562 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" podStartSLOduration=64.574551241 podStartE2EDuration="1m4.574551241s" podCreationTimestamp="2026-02-28 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:56.572539362 +0000 UTC m=+133.484547100" watchObservedRunningTime="2026-02-28 10:37:56.574551241 +0000 UTC m=+133.486558979" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.590485 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" event={"ID":"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d","Type":"ContainerStarted","Data":"3c3b558124be129f64b3ee5080bd915b33e7e76abeb55de2b0895cceb06683dc"} Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.590941 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.590976 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.599591 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.604522 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.706619 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.805541 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" podStartSLOduration=5.805518263 podStartE2EDuration="5.805518263s" podCreationTimestamp="2026-02-28 10:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:56.798588087 +0000 UTC m=+133.710595825" watchObservedRunningTime="2026-02-28 10:37:56.805518263 +0000 UTC m=+133.717525991" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.813047 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:56 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:56 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:56 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.813118 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:56 crc kubenswrapper[4972]: I0228 10:37:56.916273 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jnb9w" podStartSLOduration=16.916244347 podStartE2EDuration="16.916244347s" podCreationTimestamp="2026-02-28 10:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:56.873107308 +0000 UTC m=+133.785115046" watchObservedRunningTime="2026-02-28 10:37:56.916244347 +0000 UTC m=+133.828252085" Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.645486 4972 generic.go:334] "Generic (PLEG): container finished" podID="2e05a811-1a21-48f9-899c-e1f172328e48" containerID="3c0d08f70a28c5840f1847b06f862ff49868a7aa8b968ea036e87c7697c1375f" exitCode=0 Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.660952 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerDied","Data":"3c0d08f70a28c5840f1847b06f862ff49868a7aa8b968ea036e87c7697c1375f"} Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.661055 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerStarted","Data":"219b6b33525b652596a38b34da0e8076e4e872d2430c16050b677299adbfac42"} Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.694352 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" podStartSLOduration=6.694335909 podStartE2EDuration="6.694335909s" podCreationTimestamp="2026-02-28 10:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:37:56.918935177 +0000 UTC m=+133.830942915" watchObservedRunningTime="2026-02-28 10:37:57.694335909 +0000 UTC m=+134.606343647" Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.776971 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.810422 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:57 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:57 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:57 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.810505 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:57 crc kubenswrapper[4972]: I0228 10:37:57.819037 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2xkbh" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.231718 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.402301 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir\") pod \"84685205-887c-4e3b-8c9a-6951c278e2d6\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.402502 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access\") pod \"84685205-887c-4e3b-8c9a-6951c278e2d6\" (UID: \"84685205-887c-4e3b-8c9a-6951c278e2d6\") " Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.403848 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "84685205-887c-4e3b-8c9a-6951c278e2d6" (UID: "84685205-887c-4e3b-8c9a-6951c278e2d6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.424669 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "84685205-887c-4e3b-8c9a-6951c278e2d6" (UID: "84685205-887c-4e3b-8c9a-6951c278e2d6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.503962 4972 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84685205-887c-4e3b-8c9a-6951c278e2d6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.503998 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84685205-887c-4e3b-8c9a-6951c278e2d6-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.542832 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58714: no serving certificate available for the kubelet" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.602866 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zxcst" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.689421 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.692014 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"84685205-887c-4e3b-8c9a-6951c278e2d6","Type":"ContainerDied","Data":"9f57d799b1ca6a4657f173f515ddb48684309a72c582c93649eb6d37fb3d7398"} Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.692058 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f57d799b1ca6a4657f173f515ddb48684309a72c582c93649eb6d37fb3d7398" Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.810500 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:58 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:58 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:58 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:58 crc kubenswrapper[4972]: I0228 10:37:58.810567 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:37:59 crc kubenswrapper[4972]: I0228 10:37:59.047649 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:37:59 crc kubenswrapper[4972]: I0228 10:37:59.492043 4972 ???:1] "http: TLS handshake error from 192.168.126.11:58722: no serving certificate available for the kubelet" Feb 28 10:37:59 crc kubenswrapper[4972]: I0228 10:37:59.813308 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:37:59 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:37:59 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:37:59 crc kubenswrapper[4972]: healthz check failed Feb 28 10:37:59 crc kubenswrapper[4972]: I0228 10:37:59.813399 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.142288 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537918-nvtxp"] Feb 28 10:38:00 crc kubenswrapper[4972]: E0228 10:38:00.142570 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84685205-887c-4e3b-8c9a-6951c278e2d6" containerName="pruner" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.142583 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="84685205-887c-4e3b-8c9a-6951c278e2d6" containerName="pruner" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.142692 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="84685205-887c-4e3b-8c9a-6951c278e2d6" containerName="pruner" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.143048 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537918-nvtxp"] Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.143135 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.145959 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.146015 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.146184 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.235617 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2pwj\" (UniqueName: \"kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj\") pod \"auto-csr-approver-29537918-nvtxp\" (UID: \"36f27dc9-eb59-4aef-92e5-1267b61d4fad\") " pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.336976 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2pwj\" (UniqueName: \"kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj\") pod \"auto-csr-approver-29537918-nvtxp\" (UID: \"36f27dc9-eb59-4aef-92e5-1267b61d4fad\") " pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.357705 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2pwj\" (UniqueName: \"kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj\") pod \"auto-csr-approver-29537918-nvtxp\" (UID: \"36f27dc9-eb59-4aef-92e5-1267b61d4fad\") " pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.463192 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.814206 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:00 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:38:00 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:00 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.814293 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:00 crc kubenswrapper[4972]: I0228 10:38:00.903125 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537918-nvtxp"] Feb 28 10:38:00 crc kubenswrapper[4972]: W0228 10:38:00.916335 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36f27dc9_eb59_4aef_92e5_1267b61d4fad.slice/crio-e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11 WatchSource:0}: Error finding container e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11: Status 404 returned error can't find the container with id e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11 Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.758407 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" event={"ID":"36f27dc9-eb59-4aef-92e5-1267b61d4fad","Type":"ContainerStarted","Data":"e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11"} Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.812704 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:01 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:38:01 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:01 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.812794 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.817787 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.947525 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.947558 4972 patch_prober.go:28] interesting pod/downloads-7954f5f757-mfrtc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.947585 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:38:01 crc kubenswrapper[4972]: I0228 10:38:01.947631 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mfrtc" podUID="3fe6c3de-a3ed-4315-bcb1-8dc82ebebf53" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Feb 28 10:38:02 crc kubenswrapper[4972]: I0228 10:38:02.248472 4972 patch_prober.go:28] interesting pod/console-f9d7485db-czzdx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 28 10:38:02 crc kubenswrapper[4972]: I0228 10:38:02.248536 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-czzdx" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 28 10:38:02 crc kubenswrapper[4972]: I0228 10:38:02.808935 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:02 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:38:02 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:02 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:02 crc kubenswrapper[4972]: I0228 10:38:02.809015 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:03 crc kubenswrapper[4972]: E0228 10:38:03.586602 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:03 crc kubenswrapper[4972]: E0228 10:38:03.608019 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:03 crc kubenswrapper[4972]: E0228 10:38:03.623706 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:03 crc kubenswrapper[4972]: E0228 10:38:03.623776 4972 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:03 crc kubenswrapper[4972]: I0228 10:38:03.809422 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:03 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:38:03 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:03 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:03 crc kubenswrapper[4972]: I0228 10:38:03.809488 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:03 crc kubenswrapper[4972]: I0228 10:38:03.813617 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=2.813606324 podStartE2EDuration="2.813606324s" podCreationTimestamp="2026-02-28 10:38:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:03.81008405 +0000 UTC m=+140.722091788" watchObservedRunningTime="2026-02-28 10:38:03.813606324 +0000 UTC m=+140.725614062" Feb 28 10:38:03 crc kubenswrapper[4972]: I0228 10:38:03.813786 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 28 10:38:04 crc kubenswrapper[4972]: I0228 10:38:04.811298 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:04 crc kubenswrapper[4972]: [-]has-synced failed: reason withheld Feb 28 10:38:04 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:04 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:04 crc kubenswrapper[4972]: I0228 10:38:04.811366 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:05 crc kubenswrapper[4972]: I0228 10:38:05.809120 4972 patch_prober.go:28] interesting pod/router-default-5444994796-4ftt6 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 28 10:38:05 crc kubenswrapper[4972]: [+]has-synced ok Feb 28 10:38:05 crc kubenswrapper[4972]: [+]process-running ok Feb 28 10:38:05 crc kubenswrapper[4972]: healthz check failed Feb 28 10:38:05 crc kubenswrapper[4972]: I0228 10:38:05.809506 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4ftt6" podUID="8b782f84-114b-40d3-b7ba-806e1fccc04f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 28 10:38:06 crc kubenswrapper[4972]: I0228 10:38:06.809034 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:38:06 crc kubenswrapper[4972]: I0228 10:38:06.811557 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-4ftt6" Feb 28 10:38:06 crc kubenswrapper[4972]: I0228 10:38:06.884016 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=3.883981435 podStartE2EDuration="3.883981435s" podCreationTimestamp="2026-02-28 10:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:06.852134841 +0000 UTC m=+143.764142579" watchObservedRunningTime="2026-02-28 10:38:06.883981435 +0000 UTC m=+143.795989173" Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.310346 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.313982 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" containerID="cri-o://3c3b558124be129f64b3ee5080bd915b33e7e76abeb55de2b0895cceb06683dc" gracePeriod=30 Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.333612 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.334186 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" podUID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" containerName="route-controller-manager" containerID="cri-o://3592e67bc54524c45f6bdd3781ff8a3f244fb851f74357ad1d551a786e45b122" gracePeriod=30 Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.846453 4972 generic.go:334] "Generic (PLEG): container finished" podID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerID="3c3b558124be129f64b3ee5080bd915b33e7e76abeb55de2b0895cceb06683dc" exitCode=0 Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.846550 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" event={"ID":"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d","Type":"ContainerDied","Data":"3c3b558124be129f64b3ee5080bd915b33e7e76abeb55de2b0895cceb06683dc"} Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.848745 4972 generic.go:334] "Generic (PLEG): container finished" podID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" containerID="3592e67bc54524c45f6bdd3781ff8a3f244fb851f74357ad1d551a786e45b122" exitCode=0 Feb 28 10:38:10 crc kubenswrapper[4972]: I0228 10:38:10.848822 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" event={"ID":"ba81daec-9cd5-4f06-99e5-9f01366cf4bf","Type":"ContainerDied","Data":"3592e67bc54524c45f6bdd3781ff8a3f244fb851f74357ad1d551a786e45b122"} Feb 28 10:38:11 crc kubenswrapper[4972]: I0228 10:38:11.967896 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mfrtc" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.262813 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.270199 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.756812 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.782147 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config\") pod \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.782222 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert\") pod \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.782624 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr65z\" (UniqueName: \"kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z\") pod \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.782794 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca\") pod \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\" (UID: \"ba81daec-9cd5-4f06-99e5-9f01366cf4bf\") " Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.783841 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config" (OuterVolumeSpecName: "config") pod "ba81daec-9cd5-4f06-99e5-9f01366cf4bf" (UID: "ba81daec-9cd5-4f06-99e5-9f01366cf4bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.783886 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca" (OuterVolumeSpecName: "client-ca") pod "ba81daec-9cd5-4f06-99e5-9f01366cf4bf" (UID: "ba81daec-9cd5-4f06-99e5-9f01366cf4bf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.792366 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z" (OuterVolumeSpecName: "kube-api-access-wr65z") pod "ba81daec-9cd5-4f06-99e5-9f01366cf4bf" (UID: "ba81daec-9cd5-4f06-99e5-9f01366cf4bf"). InnerVolumeSpecName "kube-api-access-wr65z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.792384 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ba81daec-9cd5-4f06-99e5-9f01366cf4bf" (UID: "ba81daec-9cd5-4f06-99e5-9f01366cf4bf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.798291 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:12 crc kubenswrapper[4972]: E0228 10:38:12.798598 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" containerName="route-controller-manager" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.798620 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" containerName="route-controller-manager" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.798748 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" containerName="route-controller-manager" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.799284 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.814364 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.864861 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" event={"ID":"ba81daec-9cd5-4f06-99e5-9f01366cf4bf","Type":"ContainerDied","Data":"e7c1502771c6792973cc250a67981cd8e533e6c5991b0b14a198aca0605789eb"} Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.864922 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.864983 4972 scope.go:117] "RemoveContainer" containerID="3592e67bc54524c45f6bdd3781ff8a3f244fb851f74357ad1d551a786e45b122" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.883806 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-228fj\" (UniqueName: \"kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.883899 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.883973 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.884135 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.884443 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr65z\" (UniqueName: \"kubernetes.io/projected/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-kube-api-access-wr65z\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.884476 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.884488 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.884499 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81daec-9cd5-4f06-99e5-9f01366cf4bf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.902181 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.907433 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74fbb5c645-9rjwr"] Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.927896 4972 patch_prober.go:28] interesting pod/controller-manager-6487d4cd87-6gz5v container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.928203 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.985768 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-228fj\" (UniqueName: \"kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.985823 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.985864 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.985896 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.986812 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.988107 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:12 crc kubenswrapper[4972]: I0228 10:38:12.990648 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:13 crc kubenswrapper[4972]: I0228 10:38:13.010030 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-228fj\" (UniqueName: \"kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj\") pod \"route-controller-manager-845dbd6559-qpshq\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:13 crc kubenswrapper[4972]: I0228 10:38:13.145031 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:13 crc kubenswrapper[4972]: E0228 10:38:13.584996 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:13 crc kubenswrapper[4972]: E0228 10:38:13.587683 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:13 crc kubenswrapper[4972]: E0228 10:38:13.590567 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:13 crc kubenswrapper[4972]: E0228 10:38:13.590674 4972 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:13 crc kubenswrapper[4972]: I0228 10:38:13.799241 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba81daec-9cd5-4f06-99e5-9f01366cf4bf" path="/var/lib/kubelet/pods/ba81daec-9cd5-4f06-99e5-9f01366cf4bf/volumes" Feb 28 10:38:15 crc kubenswrapper[4972]: I0228 10:38:15.240762 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:38:20 crc kubenswrapper[4972]: I0228 10:38:20.000705 4972 ???:1] "http: TLS handshake error from 192.168.126.11:34384: no serving certificate available for the kubelet" Feb 28 10:38:22 crc kubenswrapper[4972]: I0228 10:38:22.900446 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2fwv6" Feb 28 10:38:23 crc kubenswrapper[4972]: E0228 10:38:23.582646 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:23 crc kubenswrapper[4972]: E0228 10:38:23.585213 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:23 crc kubenswrapper[4972]: E0228 10:38:23.586802 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:23 crc kubenswrapper[4972]: E0228 10:38:23.586886 4972 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:23 crc kubenswrapper[4972]: I0228 10:38:23.927713 4972 patch_prober.go:28] interesting pod/controller-manager-6487d4cd87-6gz5v container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:38:23 crc kubenswrapper[4972]: I0228 10:38:23.927782 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:38:25 crc kubenswrapper[4972]: I0228 10:38:25.107771 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 28 10:38:26 crc kubenswrapper[4972]: I0228 10:38:26.987274 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-5plg9_1efe1619-53ab-4d7b-83d8-c772875cf078/kube-multus-additional-cni-plugins/0.log" Feb 28 10:38:26 crc kubenswrapper[4972]: I0228 10:38:26.987806 4972 generic.go:334] "Generic (PLEG): container finished" podID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" exitCode=137 Feb 28 10:38:26 crc kubenswrapper[4972]: I0228 10:38:26.987976 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" event={"ID":"1efe1619-53ab-4d7b-83d8-c772875cf078","Type":"ContainerDied","Data":"0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d"} Feb 28 10:38:27 crc kubenswrapper[4972]: I0228 10:38:27.977247 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 28 10:38:27 crc kubenswrapper[4972]: I0228 10:38:27.978303 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:27 crc kubenswrapper[4972]: I0228 10:38:27.982272 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 28 10:38:27 crc kubenswrapper[4972]: I0228 10:38:27.982689 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 28 10:38:27 crc kubenswrapper[4972]: I0228 10:38:27.993897 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.080791 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.081430 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.183534 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.183667 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.183817 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.219555 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:28 crc kubenswrapper[4972]: I0228 10:38:28.317975 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:38:30 crc kubenswrapper[4972]: I0228 10:38:30.431597 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.775237 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.775956 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.786247 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.925245 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.925295 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:32 crc kubenswrapper[4972]: I0228 10:38:32.925358 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.027565 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.027645 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.027702 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.027720 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.027811 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.048077 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access\") pod \"installer-9-crc\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.182246 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:38:33 crc kubenswrapper[4972]: E0228 10:38:33.581276 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:33 crc kubenswrapper[4972]: E0228 10:38:33.581847 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:33 crc kubenswrapper[4972]: E0228 10:38:33.582185 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:33 crc kubenswrapper[4972]: E0228 10:38:33.582239 4972 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.927736 4972 patch_prober.go:28] interesting pod/controller-manager-6487d4cd87-6gz5v container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 28 10:38:33 crc kubenswrapper[4972]: I0228 10:38:33.927822 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.916622 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.923322 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5wpb\" (UniqueName: \"kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb\") pod \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.923624 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config\") pod \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.923713 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca\") pod \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.924105 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert\") pod \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.924218 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles\") pod \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\" (UID: \"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d\") " Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.927427 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" (UID: "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.927400 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca" (OuterVolumeSpecName: "client-ca") pod "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" (UID: "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.928543 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config" (OuterVolumeSpecName: "config") pod "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" (UID: "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.935583 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" (UID: "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.937069 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb" (OuterVolumeSpecName: "kube-api-access-z5wpb") pod "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" (UID: "1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d"). InnerVolumeSpecName "kube-api-access-z5wpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.956095 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:38:38 crc kubenswrapper[4972]: E0228 10:38:38.957292 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.957381 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.958024 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" containerName="controller-manager" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.959284 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:38 crc kubenswrapper[4972]: I0228 10:38:38.965912 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.033492 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.033773 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km5xh\" (UniqueName: \"kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034018 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034158 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034232 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034418 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034449 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034489 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034502 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.034520 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5wpb\" (UniqueName: \"kubernetes.io/projected/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d-kube-api-access-z5wpb\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.114815 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" event={"ID":"1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d","Type":"ContainerDied","Data":"5fe16a52af070fda0971e1a4044dc0bd9ec3f3043226fbee05b02e279b221d78"} Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.114943 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6487d4cd87-6gz5v" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.135054 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.135797 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.136090 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km5xh\" (UniqueName: \"kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.136129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.136216 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.137400 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.139304 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.140798 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.152430 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.157338 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6487d4cd87-6gz5v"] Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.158745 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km5xh\" (UniqueName: \"kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.190206 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca\") pod \"controller-manager-5b95795899-cvf2j\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.303222 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:39 crc kubenswrapper[4972]: I0228 10:38:39.802867 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d" path="/var/lib/kubelet/pods/1dfbbbe2-7757-4ea4-88d3-cef4d12deb6d/volumes" Feb 28 10:38:40 crc kubenswrapper[4972]: E0228 10:38:40.987609 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 28 10:38:40 crc kubenswrapper[4972]: E0228 10:38:40.988095 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-85kkr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dqghp_openshift-marketplace(0bade0e0-1dde-4239-9afd-a4cfc285eecb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:40 crc kubenswrapper[4972]: E0228 10:38:40.993004 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dqghp" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" Feb 28 10:38:43 crc kubenswrapper[4972]: E0228 10:38:43.581502 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:43 crc kubenswrapper[4972]: E0228 10:38:43.582218 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:43 crc kubenswrapper[4972]: E0228 10:38:43.582946 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:43 crc kubenswrapper[4972]: E0228 10:38:43.583054 4972 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:44 crc kubenswrapper[4972]: E0228 10:38:44.418477 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dqghp" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" Feb 28 10:38:45 crc kubenswrapper[4972]: I0228 10:38:45.400684 4972 scope.go:117] "RemoveContainer" containerID="3c3b558124be129f64b3ee5080bd915b33e7e76abeb55de2b0895cceb06683dc" Feb 28 10:38:45 crc kubenswrapper[4972]: E0228 10:38:45.709404 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 28 10:38:45 crc kubenswrapper[4972]: E0228 10:38:45.709875 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgw7r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wzx7z_openshift-marketplace(285bfdff-92e3-4ec5-bbd8-745860eef1fa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:45 crc kubenswrapper[4972]: E0228 10:38:45.712413 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wzx7z" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" Feb 28 10:38:51 crc kubenswrapper[4972]: E0228 10:38:51.372014 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wzx7z" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" Feb 28 10:38:51 crc kubenswrapper[4972]: E0228 10:38:51.482186 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 28 10:38:51 crc kubenswrapper[4972]: E0228 10:38:51.482779 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l92mf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w8pgq_openshift-marketplace(2e05a811-1a21-48f9-899c-e1f172328e48): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:51 crc kubenswrapper[4972]: E0228 10:38:51.483987 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w8pgq" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.581374 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.583098 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.584003 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.584052 4972 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.690972 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w8pgq" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.988979 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.989230 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hzjp6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2j69x_openshift-marketplace(ad7908e4-f813-4c17-a672-a9ff32bdfd8c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:53 crc kubenswrapper[4972]: E0228 10:38:53.990446 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2j69x" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" Feb 28 10:38:54 crc kubenswrapper[4972]: E0228 10:38:54.372986 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 28 10:38:54 crc kubenswrapper[4972]: E0228 10:38:54.373169 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m9dq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bccmr_openshift-marketplace(c3023e1b-7bcf-46b1-b61c-92cf0251647e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:54 crc kubenswrapper[4972]: E0228 10:38:54.374361 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bccmr" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" Feb 28 10:38:55 crc kubenswrapper[4972]: E0228 10:38:55.577098 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bccmr" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" Feb 28 10:38:55 crc kubenswrapper[4972]: E0228 10:38:55.577135 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2j69x" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" Feb 28 10:38:55 crc kubenswrapper[4972]: E0228 10:38:55.640948 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 28 10:38:55 crc kubenswrapper[4972]: E0228 10:38:55.641231 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mgpd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6jglk_openshift-marketplace(66ead19f-03ec-41d1-82b4-9697bbb1e6a6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:55 crc kubenswrapper[4972]: E0228 10:38:55.642542 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6jglk" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.508451 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6jglk" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.547497 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.547696 4972 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 28 10:38:56 crc kubenswrapper[4972]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 28 10:38:56 crc kubenswrapper[4972]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f2pwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29537918-nvtxp_openshift-infra(36f27dc9-eb59-4aef-92e5-1267b61d4fad): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Feb 28 10:38:56 crc kubenswrapper[4972]: > logger="UnhandledError" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.548882 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.603451 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.603500 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.603625 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rlnmm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-f45gq_openshift-marketplace(af90cea3-090d-4f18-88e1-c6de1a5fc469): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.603756 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgtk4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-85647_openshift-marketplace(5cea766f-d665-474b-a7fa-5e193ce3c1fd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.604376 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-5plg9_1efe1619-53ab-4d7b-83d8-c772875cf078/kube-multus-additional-cni-plugins/0.log" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.604571 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.604790 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-f45gq" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" Feb 28 10:38:56 crc kubenswrapper[4972]: E0228 10:38:56.604902 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-85647" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.693906 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist\") pod \"1efe1619-53ab-4d7b-83d8-c772875cf078\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.694418 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir\") pod \"1efe1619-53ab-4d7b-83d8-c772875cf078\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.694453 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qdjk\" (UniqueName: \"kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk\") pod \"1efe1619-53ab-4d7b-83d8-c772875cf078\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.694567 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready\") pod \"1efe1619-53ab-4d7b-83d8-c772875cf078\" (UID: \"1efe1619-53ab-4d7b-83d8-c772875cf078\") " Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.694660 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "1efe1619-53ab-4d7b-83d8-c772875cf078" (UID: "1efe1619-53ab-4d7b-83d8-c772875cf078"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.694864 4972 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1efe1619-53ab-4d7b-83d8-c772875cf078-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.695479 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready" (OuterVolumeSpecName: "ready") pod "1efe1619-53ab-4d7b-83d8-c772875cf078" (UID: "1efe1619-53ab-4d7b-83d8-c772875cf078"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.695620 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "1efe1619-53ab-4d7b-83d8-c772875cf078" (UID: "1efe1619-53ab-4d7b-83d8-c772875cf078"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.717149 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk" (OuterVolumeSpecName: "kube-api-access-4qdjk") pod "1efe1619-53ab-4d7b-83d8-c772875cf078" (UID: "1efe1619-53ab-4d7b-83d8-c772875cf078"). InnerVolumeSpecName "kube-api-access-4qdjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.795882 4972 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1efe1619-53ab-4d7b-83d8-c772875cf078-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.795938 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qdjk\" (UniqueName: \"kubernetes.io/projected/1efe1619-53ab-4d7b-83d8-c772875cf078-kube-api-access-4qdjk\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:56 crc kubenswrapper[4972]: I0228 10:38:56.795956 4972 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/1efe1619-53ab-4d7b-83d8-c772875cf078-ready\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.018721 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.036417 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 28 10:38:57 crc kubenswrapper[4972]: W0228 10:38:57.054854 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd26eeeef_20d2_43c9_b132_9699eb84f9a4.slice/crio-7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7 WatchSource:0}: Error finding container 7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7: Status 404 returned error can't find the container with id 7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7 Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.119861 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.145403 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 28 10:38:57 crc kubenswrapper[4972]: W0228 10:38:57.156791 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc728f38a_b1ad_4ce3_a3c4_a4a82596e70d.slice/crio-c62f349dd8480e91ad7fa94eb67665c16381e705ee77c43cc0b23ffbbfe83d5a WatchSource:0}: Error finding container c62f349dd8480e91ad7fa94eb67665c16381e705ee77c43cc0b23ffbbfe83d5a: Status 404 returned error can't find the container with id c62f349dd8480e91ad7fa94eb67665c16381e705ee77c43cc0b23ffbbfe83d5a Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.233474 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" event={"ID":"ea630ded-10a5-4389-a876-22488dd3a0de","Type":"ContainerStarted","Data":"8fdc8a6da196bc0df9877aadb823c58bc2c6922fd0a9d8ea8399c5e8d1b23623"} Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.235282 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d26eeeef-20d2-43c9-b132-9699eb84f9a4","Type":"ContainerStarted","Data":"7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7"} Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.237429 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" event={"ID":"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d","Type":"ContainerStarted","Data":"c62f349dd8480e91ad7fa94eb67665c16381e705ee77c43cc0b23ffbbfe83d5a"} Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.238388 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70bfef93-4319-4688-9760-f67d973ec35c","Type":"ContainerStarted","Data":"89305d83ba71bb98e6bcd3556b007b8aed28f076223142c354d00a20616db7b0"} Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.243018 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-5plg9_1efe1619-53ab-4d7b-83d8-c772875cf078/kube-multus-additional-cni-plugins/0.log" Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.243259 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" event={"ID":"1efe1619-53ab-4d7b-83d8-c772875cf078","Type":"ContainerDied","Data":"dc35327ec7148954bef1417a3e030fb2da3dfb3643bf129c2fbdd527261baf5e"} Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.243479 4972 scope.go:117] "RemoveContainer" containerID="0be7a03f3dca401c019779630511c7019eec14d8e2051f4f0008a3365ffdbf9d" Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.243956 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-5plg9" Feb 28 10:38:57 crc kubenswrapper[4972]: E0228 10:38:57.246428 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-85647" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" Feb 28 10:38:57 crc kubenswrapper[4972]: E0228 10:38:57.249327 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" Feb 28 10:38:57 crc kubenswrapper[4972]: E0228 10:38:57.250251 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-f45gq" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.351446 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-5plg9"] Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.356207 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-5plg9"] Feb 28 10:38:57 crc kubenswrapper[4972]: I0228 10:38:57.798026 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" path="/var/lib/kubelet/pods/1efe1619-53ab-4d7b-83d8-c772875cf078/volumes" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.258702 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" event={"ID":"ea630ded-10a5-4389-a876-22488dd3a0de","Type":"ContainerStarted","Data":"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843"} Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.258933 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" podUID="ea630ded-10a5-4389-a876-22488dd3a0de" containerName="route-controller-manager" containerID="cri-o://f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843" gracePeriod=30 Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.259195 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.261271 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d26eeeef-20d2-43c9-b132-9699eb84f9a4","Type":"ContainerStarted","Data":"e695cbcf8a37b860dc918b4cea879365e51f1ebe111a288da4ef7c32e4b97281"} Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.265595 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" event={"ID":"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d","Type":"ContainerStarted","Data":"8bfd915be868a8ace030a1d58dafe37d8c36dd2c77b375c6cea626a6247c6ef1"} Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.266421 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.269665 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.270266 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70bfef93-4319-4688-9760-f67d973ec35c","Type":"ContainerStarted","Data":"c99513ea4524df530d7a098f055aaab73c10a33486a14121c81415980a6c91ef"} Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.270920 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.285842 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" podStartSLOduration=48.285772312 podStartE2EDuration="48.285772312s" podCreationTimestamp="2026-02-28 10:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:58.281860275 +0000 UTC m=+195.193868023" watchObservedRunningTime="2026-02-28 10:38:58.285772312 +0000 UTC m=+195.197780090" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.314514 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" podStartSLOduration=28.314497916 podStartE2EDuration="28.314497916s" podCreationTimestamp="2026-02-28 10:38:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:58.308660915 +0000 UTC m=+195.220668653" watchObservedRunningTime="2026-02-28 10:38:58.314497916 +0000 UTC m=+195.226505654" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.367791 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=26.367774378 podStartE2EDuration="26.367774378s" podCreationTimestamp="2026-02-28 10:38:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:58.365227998 +0000 UTC m=+195.277235746" watchObservedRunningTime="2026-02-28 10:38:58.367774378 +0000 UTC m=+195.279782116" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.396609 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=31.396594835 podStartE2EDuration="31.396594835s" podCreationTimestamp="2026-02-28 10:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:38:58.395146864 +0000 UTC m=+195.307154602" watchObservedRunningTime="2026-02-28 10:38:58.396594835 +0000 UTC m=+195.308602573" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.701346 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.723627 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-228fj\" (UniqueName: \"kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj\") pod \"ea630ded-10a5-4389-a876-22488dd3a0de\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.724388 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca\") pod \"ea630ded-10a5-4389-a876-22488dd3a0de\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.725169 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert\") pod \"ea630ded-10a5-4389-a876-22488dd3a0de\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.725264 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca" (OuterVolumeSpecName: "client-ca") pod "ea630ded-10a5-4389-a876-22488dd3a0de" (UID: "ea630ded-10a5-4389-a876-22488dd3a0de"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.725274 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config\") pod \"ea630ded-10a5-4389-a876-22488dd3a0de\" (UID: \"ea630ded-10a5-4389-a876-22488dd3a0de\") " Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.725625 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.726332 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config" (OuterVolumeSpecName: "config") pod "ea630ded-10a5-4389-a876-22488dd3a0de" (UID: "ea630ded-10a5-4389-a876-22488dd3a0de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.736867 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj" (OuterVolumeSpecName: "kube-api-access-228fj") pod "ea630ded-10a5-4389-a876-22488dd3a0de" (UID: "ea630ded-10a5-4389-a876-22488dd3a0de"). InnerVolumeSpecName "kube-api-access-228fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.737251 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ea630ded-10a5-4389-a876-22488dd3a0de" (UID: "ea630ded-10a5-4389-a876-22488dd3a0de"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.826986 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea630ded-10a5-4389-a876-22488dd3a0de-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.827044 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-228fj\" (UniqueName: \"kubernetes.io/projected/ea630ded-10a5-4389-a876-22488dd3a0de-kube-api-access-228fj\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:58 crc kubenswrapper[4972]: I0228 10:38:58.827061 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea630ded-10a5-4389-a876-22488dd3a0de-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.282150 4972 generic.go:334] "Generic (PLEG): container finished" podID="ea630ded-10a5-4389-a876-22488dd3a0de" containerID="f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843" exitCode=0 Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.282604 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" event={"ID":"ea630ded-10a5-4389-a876-22488dd3a0de","Type":"ContainerDied","Data":"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843"} Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.282643 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" event={"ID":"ea630ded-10a5-4389-a876-22488dd3a0de","Type":"ContainerDied","Data":"8fdc8a6da196bc0df9877aadb823c58bc2c6922fd0a9d8ea8399c5e8d1b23623"} Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.282668 4972 scope.go:117] "RemoveContainer" containerID="f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843" Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.282817 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq" Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.291780 4972 generic.go:334] "Generic (PLEG): container finished" podID="70bfef93-4319-4688-9760-f67d973ec35c" containerID="c99513ea4524df530d7a098f055aaab73c10a33486a14121c81415980a6c91ef" exitCode=0 Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.291868 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70bfef93-4319-4688-9760-f67d973ec35c","Type":"ContainerDied","Data":"c99513ea4524df530d7a098f055aaab73c10a33486a14121c81415980a6c91ef"} Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.341076 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.344214 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-845dbd6559-qpshq"] Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.355502 4972 scope.go:117] "RemoveContainer" containerID="f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843" Feb 28 10:38:59 crc kubenswrapper[4972]: E0228 10:38:59.356277 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843\": container with ID starting with f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843 not found: ID does not exist" containerID="f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843" Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.356380 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843"} err="failed to get container status \"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843\": rpc error: code = NotFound desc = could not find container \"f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843\": container with ID starting with f7189d59132b664ae4f106dc4418b08ed36e724039ae236b3d8c3ba2b06b1843 not found: ID does not exist" Feb 28 10:38:59 crc kubenswrapper[4972]: I0228 10:38:59.798237 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea630ded-10a5-4389-a876-22488dd3a0de" path="/var/lib/kubelet/pods/ea630ded-10a5-4389-a876-22488dd3a0de/volumes" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.302697 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerID="06e657f11c83ddaaeeb67cf72e76f6b437469dad1f9f3897258650ef8527ef82" exitCode=0 Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.302814 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerDied","Data":"06e657f11c83ddaaeeb67cf72e76f6b437469dad1f9f3897258650ef8527ef82"} Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.641958 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.666440 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir\") pod \"70bfef93-4319-4688-9760-f67d973ec35c\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.666545 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access\") pod \"70bfef93-4319-4688-9760-f67d973ec35c\" (UID: \"70bfef93-4319-4688-9760-f67d973ec35c\") " Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.666563 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "70bfef93-4319-4688-9760-f67d973ec35c" (UID: "70bfef93-4319-4688-9760-f67d973ec35c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.666784 4972 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70bfef93-4319-4688-9760-f67d973ec35c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.677804 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "70bfef93-4319-4688-9760-f67d973ec35c" (UID: "70bfef93-4319-4688-9760-f67d973ec35c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748166 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:00 crc kubenswrapper[4972]: E0228 10:39:00.748542 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea630ded-10a5-4389-a876-22488dd3a0de" containerName="route-controller-manager" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748570 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea630ded-10a5-4389-a876-22488dd3a0de" containerName="route-controller-manager" Feb 28 10:39:00 crc kubenswrapper[4972]: E0228 10:39:00.748594 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748605 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:39:00 crc kubenswrapper[4972]: E0228 10:39:00.748635 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bfef93-4319-4688-9760-f67d973ec35c" containerName="pruner" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748644 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bfef93-4319-4688-9760-f67d973ec35c" containerName="pruner" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748775 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="1efe1619-53ab-4d7b-83d8-c772875cf078" containerName="kube-multus-additional-cni-plugins" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748795 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bfef93-4319-4688-9760-f67d973ec35c" containerName="pruner" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.748804 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea630ded-10a5-4389-a876-22488dd3a0de" containerName="route-controller-manager" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.749790 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.753961 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.754663 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.754881 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.755042 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.755108 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.754999 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.762940 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.775794 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.775940 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.775974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.776067 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb88l\" (UniqueName: \"kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.776214 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70bfef93-4319-4688-9760-f67d973ec35c-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.878110 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.879505 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.879616 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.879782 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb88l\" (UniqueName: \"kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.880090 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.885168 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.890558 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.904072 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb88l\" (UniqueName: \"kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l\") pod \"route-controller-manager-66bdc57965-fxhd5\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:00 crc kubenswrapper[4972]: I0228 10:39:00.998185 4972 ???:1] "http: TLS handshake error from 192.168.126.11:33024: no serving certificate available for the kubelet" Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.116325 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.316367 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.316369 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"70bfef93-4319-4688-9760-f67d973ec35c","Type":"ContainerDied","Data":"89305d83ba71bb98e6bcd3556b007b8aed28f076223142c354d00a20616db7b0"} Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.316808 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89305d83ba71bb98e6bcd3556b007b8aed28f076223142c354d00a20616db7b0" Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.320893 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerStarted","Data":"a87faa3ec6ec59670c6dc44deed17678a472cdc8b758a80f31a7d5dfcce24ec4"} Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.548727 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dqghp" podStartSLOduration=5.342456146 podStartE2EDuration="1m9.548695232s" podCreationTimestamp="2026-02-28 10:37:52 +0000 UTC" firstStartedPulling="2026-02-28 10:37:56.507874884 +0000 UTC m=+133.419882622" lastFinishedPulling="2026-02-28 10:39:00.71411396 +0000 UTC m=+197.626121708" observedRunningTime="2026-02-28 10:39:01.345487386 +0000 UTC m=+198.257495124" watchObservedRunningTime="2026-02-28 10:39:01.548695232 +0000 UTC m=+198.460702970" Feb 28 10:39:01 crc kubenswrapper[4972]: I0228 10:39:01.550051 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:01 crc kubenswrapper[4972]: W0228 10:39:01.555447 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ecf0af8_b7a9_4489_b848_51ab6dfcc1b4.slice/crio-81bdb0e2474028fd03939625a49550d12278c15487f93d7bd77e72d0b5fb7431 WatchSource:0}: Error finding container 81bdb0e2474028fd03939625a49550d12278c15487f93d7bd77e72d0b5fb7431: Status 404 returned error can't find the container with id 81bdb0e2474028fd03939625a49550d12278c15487f93d7bd77e72d0b5fb7431 Feb 28 10:39:02 crc kubenswrapper[4972]: I0228 10:39:02.330189 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" event={"ID":"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4","Type":"ContainerStarted","Data":"33cd4d8e80b487c5110ba2d646255f3def605f663411ee2abce815574f19ebb8"} Feb 28 10:39:02 crc kubenswrapper[4972]: I0228 10:39:02.330767 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" event={"ID":"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4","Type":"ContainerStarted","Data":"81bdb0e2474028fd03939625a49550d12278c15487f93d7bd77e72d0b5fb7431"} Feb 28 10:39:02 crc kubenswrapper[4972]: I0228 10:39:02.330825 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:02 crc kubenswrapper[4972]: I0228 10:39:02.352657 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:02 crc kubenswrapper[4972]: I0228 10:39:02.375428 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" podStartSLOduration=32.375403453 podStartE2EDuration="32.375403453s" podCreationTimestamp="2026-02-28 10:38:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:39:02.352832801 +0000 UTC m=+199.264840539" watchObservedRunningTime="2026-02-28 10:39:02.375403453 +0000 UTC m=+199.287411191" Feb 28 10:39:03 crc kubenswrapper[4972]: I0228 10:39:03.021093 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:03 crc kubenswrapper[4972]: I0228 10:39:03.021293 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:04 crc kubenswrapper[4972]: I0228 10:39:04.243062 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dqghp" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="registry-server" probeResult="failure" output=< Feb 28 10:39:04 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:39:04 crc kubenswrapper[4972]: > Feb 28 10:39:06 crc kubenswrapper[4972]: I0228 10:39:06.370042 4972 generic.go:334] "Generic (PLEG): container finished" podID="2e05a811-1a21-48f9-899c-e1f172328e48" containerID="ce3c57909f4ec6b5f1fc4e0a0fdb6bb2c428b3fec423c0723f83d8eb8aff6f54" exitCode=0 Feb 28 10:39:06 crc kubenswrapper[4972]: I0228 10:39:06.370268 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerDied","Data":"ce3c57909f4ec6b5f1fc4e0a0fdb6bb2c428b3fec423c0723f83d8eb8aff6f54"} Feb 28 10:39:07 crc kubenswrapper[4972]: I0228 10:39:07.380628 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerStarted","Data":"47d0b5bcd569024b1786c88074ecd3b7706bc846fef060cc6552e86f51a7151d"} Feb 28 10:39:07 crc kubenswrapper[4972]: I0228 10:39:07.384036 4972 generic.go:334] "Generic (PLEG): container finished" podID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerID="8187d38cae4b15655eb3003e03926b8cc107a474bc480a2198928a802dd7dc0c" exitCode=0 Feb 28 10:39:07 crc kubenswrapper[4972]: I0228 10:39:07.384103 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerDied","Data":"8187d38cae4b15655eb3003e03926b8cc107a474bc480a2198928a802dd7dc0c"} Feb 28 10:39:07 crc kubenswrapper[4972]: I0228 10:39:07.412916 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w8pgq" podStartSLOduration=3.271140276 podStartE2EDuration="1m12.41288368s" podCreationTimestamp="2026-02-28 10:37:55 +0000 UTC" firstStartedPulling="2026-02-28 10:37:57.669559424 +0000 UTC m=+134.581567162" lastFinishedPulling="2026-02-28 10:39:06.811302828 +0000 UTC m=+203.723310566" observedRunningTime="2026-02-28 10:39:07.410442502 +0000 UTC m=+204.322450240" watchObservedRunningTime="2026-02-28 10:39:07.41288368 +0000 UTC m=+204.324891418" Feb 28 10:39:08 crc kubenswrapper[4972]: I0228 10:39:08.391424 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerStarted","Data":"42a51a97831116a449d0f1dd1d0d035feafd9a2fb8a851a26796448131519a78"} Feb 28 10:39:08 crc kubenswrapper[4972]: I0228 10:39:08.814199 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wzx7z" podStartSLOduration=3.2905691839999998 podStartE2EDuration="1m16.814177708s" podCreationTimestamp="2026-02-28 10:37:52 +0000 UTC" firstStartedPulling="2026-02-28 10:37:54.418603987 +0000 UTC m=+131.330611725" lastFinishedPulling="2026-02-28 10:39:07.942212511 +0000 UTC m=+204.854220249" observedRunningTime="2026-02-28 10:39:08.41073496 +0000 UTC m=+205.322742698" watchObservedRunningTime="2026-02-28 10:39:08.814177708 +0000 UTC m=+205.726185446" Feb 28 10:39:10 crc kubenswrapper[4972]: I0228 10:39:10.408236 4972 generic.go:334] "Generic (PLEG): container finished" podID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerID="f962e25bab637209452d9ce7a4762fae072d2350ff586d7d2fcf502374f44562" exitCode=0 Feb 28 10:39:10 crc kubenswrapper[4972]: I0228 10:39:10.408315 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerDied","Data":"f962e25bab637209452d9ce7a4762fae072d2350ff586d7d2fcf502374f44562"} Feb 28 10:39:10 crc kubenswrapper[4972]: I0228 10:39:10.412548 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerStarted","Data":"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257"} Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.421662 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" event={"ID":"36f27dc9-eb59-4aef-92e5-1267b61d4fad","Type":"ContainerStarted","Data":"0bab11e9a64188cb3222a15eee1b23cb1acaf2d2bacf6bd87281b853e5ba246f"} Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.425697 4972 generic.go:334] "Generic (PLEG): container finished" podID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerID="3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257" exitCode=0 Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.425783 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerDied","Data":"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257"} Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.440818 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" podStartSLOduration=1.763951337 podStartE2EDuration="1m11.4407879s" podCreationTimestamp="2026-02-28 10:38:00 +0000 UTC" firstStartedPulling="2026-02-28 10:38:00.923895053 +0000 UTC m=+137.835902801" lastFinishedPulling="2026-02-28 10:39:10.600731626 +0000 UTC m=+207.512739364" observedRunningTime="2026-02-28 10:39:11.439199566 +0000 UTC m=+208.351207304" watchObservedRunningTime="2026-02-28 10:39:11.4407879 +0000 UTC m=+208.352795638" Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.893974 4972 csr.go:261] certificate signing request csr-jlntc is approved, waiting to be issued Feb 28 10:39:11 crc kubenswrapper[4972]: I0228 10:39:11.901998 4972 csr.go:257] certificate signing request csr-jlntc is issued Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.432020 4972 generic.go:334] "Generic (PLEG): container finished" podID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" containerID="0bab11e9a64188cb3222a15eee1b23cb1acaf2d2bacf6bd87281b853e5ba246f" exitCode=0 Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.432187 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" event={"ID":"36f27dc9-eb59-4aef-92e5-1267b61d4fad","Type":"ContainerDied","Data":"0bab11e9a64188cb3222a15eee1b23cb1acaf2d2bacf6bd87281b853e5ba246f"} Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.435552 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerStarted","Data":"fb61785bbe1dfd1d8c33eaa49b10dec9e8523527e2940586abdb909d0bbcd54d"} Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.636066 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.636128 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.762343 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.903175 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-16 08:04:52.360925377 +0000 UTC Feb 28 10:39:12 crc kubenswrapper[4972]: I0228 10:39:12.903220 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6261h25m39.457708148s for next certificate rotation Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.025501 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.073420 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.469818 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6jglk" podStartSLOduration=4.231566533 podStartE2EDuration="1m21.469802582s" podCreationTimestamp="2026-02-28 10:37:52 +0000 UTC" firstStartedPulling="2026-02-28 10:37:54.41872037 +0000 UTC m=+131.330728108" lastFinishedPulling="2026-02-28 10:39:11.656956419 +0000 UTC m=+208.568964157" observedRunningTime="2026-02-28 10:39:13.465018868 +0000 UTC m=+210.377026606" watchObservedRunningTime="2026-02-28 10:39:13.469802582 +0000 UTC m=+210.381810320" Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.501255 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.903542 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-09 06:28:42.185786901 +0000 UTC Feb 28 10:39:13 crc kubenswrapper[4972]: I0228 10:39:13.903843 4972 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7555h49m28.281946795s for next certificate rotation Feb 28 10:39:14 crc kubenswrapper[4972]: I0228 10:39:14.448484 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerStarted","Data":"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809"} Feb 28 10:39:14 crc kubenswrapper[4972]: I0228 10:39:14.467261 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f45gq" podStartSLOduration=2.906587199 podStartE2EDuration="1m19.46723988s" podCreationTimestamp="2026-02-28 10:37:55 +0000 UTC" firstStartedPulling="2026-02-28 10:37:56.573086798 +0000 UTC m=+133.485094536" lastFinishedPulling="2026-02-28 10:39:13.133739479 +0000 UTC m=+210.045747217" observedRunningTime="2026-02-28 10:39:14.466133289 +0000 UTC m=+211.378141027" watchObservedRunningTime="2026-02-28 10:39:14.46723988 +0000 UTC m=+211.379247618" Feb 28 10:39:15 crc kubenswrapper[4972]: I0228 10:39:15.775898 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:39:15 crc kubenswrapper[4972]: I0228 10:39:15.776631 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:39:16 crc kubenswrapper[4972]: I0228 10:39:16.210275 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:16 crc kubenswrapper[4972]: I0228 10:39:16.210779 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:16 crc kubenswrapper[4972]: I0228 10:39:16.246412 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:16 crc kubenswrapper[4972]: I0228 10:39:16.500841 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:16 crc kubenswrapper[4972]: I0228 10:39:16.841910 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f45gq" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="registry-server" probeResult="failure" output=< Feb 28 10:39:16 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:39:16 crc kubenswrapper[4972]: > Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.031840 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.032234 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dqghp" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="registry-server" containerID="cri-o://a87faa3ec6ec59670c6dc44deed17678a472cdc8b758a80f31a7d5dfcce24ec4" gracePeriod=2 Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.467802 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerID="a87faa3ec6ec59670c6dc44deed17678a472cdc8b758a80f31a7d5dfcce24ec4" exitCode=0 Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.467925 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerDied","Data":"a87faa3ec6ec59670c6dc44deed17678a472cdc8b758a80f31a7d5dfcce24ec4"} Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.516743 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.641697 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2pwj\" (UniqueName: \"kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj\") pod \"36f27dc9-eb59-4aef-92e5-1267b61d4fad\" (UID: \"36f27dc9-eb59-4aef-92e5-1267b61d4fad\") " Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.649808 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj" (OuterVolumeSpecName: "kube-api-access-f2pwj") pod "36f27dc9-eb59-4aef-92e5-1267b61d4fad" (UID: "36f27dc9-eb59-4aef-92e5-1267b61d4fad"). InnerVolumeSpecName "kube-api-access-f2pwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:17 crc kubenswrapper[4972]: I0228 10:39:17.743694 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2pwj\" (UniqueName: \"kubernetes.io/projected/36f27dc9-eb59-4aef-92e5-1267b61d4fad-kube-api-access-f2pwj\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:18 crc kubenswrapper[4972]: I0228 10:39:18.026273 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:39:18 crc kubenswrapper[4972]: I0228 10:39:18.475852 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" event={"ID":"36f27dc9-eb59-4aef-92e5-1267b61d4fad","Type":"ContainerDied","Data":"e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11"} Feb 28 10:39:18 crc kubenswrapper[4972]: I0228 10:39:18.475929 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e78b3f9986188c1ae14c96fa238c2e8226dfc017fe9ee1e5fdcb61a116c62d11" Feb 28 10:39:18 crc kubenswrapper[4972]: I0228 10:39:18.476122 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w8pgq" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="registry-server" containerID="cri-o://47d0b5bcd569024b1786c88074ecd3b7706bc846fef060cc6552e86f51a7151d" gracePeriod=2 Feb 28 10:39:18 crc kubenswrapper[4972]: I0228 10:39:18.476214 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537918-nvtxp" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.486239 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dqghp" event={"ID":"0bade0e0-1dde-4239-9afd-a4cfc285eecb","Type":"ContainerDied","Data":"325083a46718fc482f039ad27d5c92a2dfad1fce3cdb5dac23930b15949b1b98"} Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.486541 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="325083a46718fc482f039ad27d5c92a2dfad1fce3cdb5dac23930b15949b1b98" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.489003 4972 generic.go:334] "Generic (PLEG): container finished" podID="2e05a811-1a21-48f9-899c-e1f172328e48" containerID="47d0b5bcd569024b1786c88074ecd3b7706bc846fef060cc6552e86f51a7151d" exitCode=0 Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.489047 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerDied","Data":"47d0b5bcd569024b1786c88074ecd3b7706bc846fef060cc6552e86f51a7151d"} Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.495422 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.670917 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85kkr\" (UniqueName: \"kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr\") pod \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.670973 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities\") pod \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.671023 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content\") pod \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\" (UID: \"0bade0e0-1dde-4239-9afd-a4cfc285eecb\") " Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.672257 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities" (OuterVolumeSpecName: "utilities") pod "0bade0e0-1dde-4239-9afd-a4cfc285eecb" (UID: "0bade0e0-1dde-4239-9afd-a4cfc285eecb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.678145 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr" (OuterVolumeSpecName: "kube-api-access-85kkr") pod "0bade0e0-1dde-4239-9afd-a4cfc285eecb" (UID: "0bade0e0-1dde-4239-9afd-a4cfc285eecb"). InnerVolumeSpecName "kube-api-access-85kkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.738701 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bade0e0-1dde-4239-9afd-a4cfc285eecb" (UID: "0bade0e0-1dde-4239-9afd-a4cfc285eecb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.772637 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85kkr\" (UniqueName: \"kubernetes.io/projected/0bade0e0-1dde-4239-9afd-a4cfc285eecb-kube-api-access-85kkr\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.772666 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.772677 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bade0e0-1dde-4239-9afd-a4cfc285eecb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:19 crc kubenswrapper[4972]: I0228 10:39:19.948647 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.075855 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l92mf\" (UniqueName: \"kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf\") pod \"2e05a811-1a21-48f9-899c-e1f172328e48\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.076241 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities\") pod \"2e05a811-1a21-48f9-899c-e1f172328e48\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.076346 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content\") pod \"2e05a811-1a21-48f9-899c-e1f172328e48\" (UID: \"2e05a811-1a21-48f9-899c-e1f172328e48\") " Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.077395 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities" (OuterVolumeSpecName: "utilities") pod "2e05a811-1a21-48f9-899c-e1f172328e48" (UID: "2e05a811-1a21-48f9-899c-e1f172328e48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.079683 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf" (OuterVolumeSpecName: "kube-api-access-l92mf") pod "2e05a811-1a21-48f9-899c-e1f172328e48" (UID: "2e05a811-1a21-48f9-899c-e1f172328e48"). InnerVolumeSpecName "kube-api-access-l92mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.177986 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l92mf\" (UniqueName: \"kubernetes.io/projected/2e05a811-1a21-48f9-899c-e1f172328e48-kube-api-access-l92mf\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.178021 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.205063 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e05a811-1a21-48f9-899c-e1f172328e48" (UID: "2e05a811-1a21-48f9-899c-e1f172328e48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.279603 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e05a811-1a21-48f9-899c-e1f172328e48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.496293 4972 generic.go:334] "Generic (PLEG): container finished" podID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerID="c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355" exitCode=0 Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.496353 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerDied","Data":"c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355"} Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.500921 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8pgq" event={"ID":"2e05a811-1a21-48f9-899c-e1f172328e48","Type":"ContainerDied","Data":"219b6b33525b652596a38b34da0e8076e4e872d2430c16050b677299adbfac42"} Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.500966 4972 scope.go:117] "RemoveContainer" containerID="47d0b5bcd569024b1786c88074ecd3b7706bc846fef060cc6552e86f51a7151d" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.501074 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8pgq" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.515248 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerID="7927e77085602432e3c47acb85b6653a07d802ba23fa800f7c891f923abd27ce" exitCode=0 Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.515398 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerDied","Data":"7927e77085602432e3c47acb85b6653a07d802ba23fa800f7c891f923abd27ce"} Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.525651 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerID="ca732736e7da1d057ab4a5a806eb7664826d03a02c4ea0418ab3cfe348325ae0" exitCode=0 Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.525748 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dqghp" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.526287 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerDied","Data":"ca732736e7da1d057ab4a5a806eb7664826d03a02c4ea0418ab3cfe348325ae0"} Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.551557 4972 scope.go:117] "RemoveContainer" containerID="ce3c57909f4ec6b5f1fc4e0a0fdb6bb2c428b3fec423c0723f83d8eb8aff6f54" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.564681 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.568420 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dqghp"] Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.581922 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.584661 4972 scope.go:117] "RemoveContainer" containerID="3c0d08f70a28c5840f1847b06f862ff49868a7aa8b968ea036e87c7697c1375f" Feb 28 10:39:20 crc kubenswrapper[4972]: I0228 10:39:20.587084 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w8pgq"] Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.536371 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerStarted","Data":"8dc96ea3620ee308b11636e4c5f6a0bde15eed6a3212b4eeffff4759fb731756"} Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.538653 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerStarted","Data":"c20c683f6acdd1ff7a3d66ed39ad63230a4acbb45c2495ec4b7e87ec4b52cb6d"} Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.540386 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerStarted","Data":"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5"} Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.559813 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2j69x" podStartSLOduration=2.934346925 podStartE2EDuration="1m27.559797438s" podCreationTimestamp="2026-02-28 10:37:54 +0000 UTC" firstStartedPulling="2026-02-28 10:37:56.435559919 +0000 UTC m=+133.347567667" lastFinishedPulling="2026-02-28 10:39:21.061010432 +0000 UTC m=+217.973018180" observedRunningTime="2026-02-28 10:39:21.555883108 +0000 UTC m=+218.467890846" watchObservedRunningTime="2026-02-28 10:39:21.559797438 +0000 UTC m=+218.471805176" Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.590048 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bccmr" podStartSLOduration=2.856075731 podStartE2EDuration="1m27.590030534s" podCreationTimestamp="2026-02-28 10:37:54 +0000 UTC" firstStartedPulling="2026-02-28 10:37:56.462484207 +0000 UTC m=+133.374491945" lastFinishedPulling="2026-02-28 10:39:21.196439 +0000 UTC m=+218.108446748" observedRunningTime="2026-02-28 10:39:21.585519387 +0000 UTC m=+218.497527135" watchObservedRunningTime="2026-02-28 10:39:21.590030534 +0000 UTC m=+218.502038262" Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.614815 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-85647" podStartSLOduration=5.243633264 podStartE2EDuration="1m29.614796367s" podCreationTimestamp="2026-02-28 10:37:52 +0000 UTC" firstStartedPulling="2026-02-28 10:37:56.527489416 +0000 UTC m=+133.439497154" lastFinishedPulling="2026-02-28 10:39:20.898652519 +0000 UTC m=+217.810660257" observedRunningTime="2026-02-28 10:39:21.610604099 +0000 UTC m=+218.522611847" watchObservedRunningTime="2026-02-28 10:39:21.614796367 +0000 UTC m=+218.526804115" Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.795328 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" path="/var/lib/kubelet/pods/0bade0e0-1dde-4239-9afd-a4cfc285eecb/volumes" Feb 28 10:39:21 crc kubenswrapper[4972]: I0228 10:39:21.796085 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" path="/var/lib/kubelet/pods/2e05a811-1a21-48f9-899c-e1f172328e48/volumes" Feb 28 10:39:22 crc kubenswrapper[4972]: I0228 10:39:22.754337 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:39:22 crc kubenswrapper[4972]: I0228 10:39:22.754402 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:39:22 crc kubenswrapper[4972]: I0228 10:39:22.805007 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:39:23 crc kubenswrapper[4972]: I0228 10:39:23.185419 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:23 crc kubenswrapper[4972]: I0228 10:39:23.185726 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:23 crc kubenswrapper[4972]: I0228 10:39:23.274256 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:23 crc kubenswrapper[4972]: I0228 10:39:23.589618 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:39:24 crc kubenswrapper[4972]: I0228 10:39:24.811717 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:39:24 crc kubenswrapper[4972]: I0228 10:39:24.811801 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:39:24 crc kubenswrapper[4972]: I0228 10:39:24.862810 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:39:25 crc kubenswrapper[4972]: I0228 10:39:25.142055 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:39:25 crc kubenswrapper[4972]: I0228 10:39:25.142120 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:39:25 crc kubenswrapper[4972]: I0228 10:39:25.204738 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:39:25 crc kubenswrapper[4972]: I0228 10:39:25.840108 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:39:25 crc kubenswrapper[4972]: I0228 10:39:25.888228 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.325088 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.325701 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" podUID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" containerName="controller-manager" containerID="cri-o://8bfd915be868a8ace030a1d58dafe37d8c36dd2c77b375c6cea626a6247c6ef1" gracePeriod=30 Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.424225 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.424493 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" podUID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" containerName="route-controller-manager" containerID="cri-o://33cd4d8e80b487c5110ba2d646255f3def605f663411ee2abce815574f19ebb8" gracePeriod=30 Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.596107 4972 generic.go:334] "Generic (PLEG): container finished" podID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" containerID="8bfd915be868a8ace030a1d58dafe37d8c36dd2c77b375c6cea626a6247c6ef1" exitCode=0 Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.596190 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" event={"ID":"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d","Type":"ContainerDied","Data":"8bfd915be868a8ace030a1d58dafe37d8c36dd2c77b375c6cea626a6247c6ef1"} Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.598032 4972 generic.go:334] "Generic (PLEG): container finished" podID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" containerID="33cd4d8e80b487c5110ba2d646255f3def605f663411ee2abce815574f19ebb8" exitCode=0 Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.598062 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" event={"ID":"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4","Type":"ContainerDied","Data":"33cd4d8e80b487c5110ba2d646255f3def605f663411ee2abce815574f19ebb8"} Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.808871 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.813570 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932794 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles\") pod \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932872 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config\") pod \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932901 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config\") pod \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932924 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert\") pod \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932947 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert\") pod \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.932993 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca\") pod \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933044 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb88l\" (UniqueName: \"kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l\") pod \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933070 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km5xh\" (UniqueName: \"kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh\") pod \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\" (UID: \"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933098 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca\") pod \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\" (UID: \"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4\") " Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933819 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" (UID: "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933841 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca" (OuterVolumeSpecName: "client-ca") pod "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" (UID: "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933885 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca" (OuterVolumeSpecName: "client-ca") pod "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" (UID: "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933914 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config" (OuterVolumeSpecName: "config") pod "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" (UID: "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.933993 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config" (OuterVolumeSpecName: "config") pod "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" (UID: "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.938092 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" (UID: "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.938156 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh" (OuterVolumeSpecName: "kube-api-access-km5xh") pod "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" (UID: "c728f38a-b1ad-4ce3-a3c4-a4a82596e70d"). InnerVolumeSpecName "kube-api-access-km5xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.938206 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l" (OuterVolumeSpecName: "kube-api-access-cb88l") pod "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" (UID: "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4"). InnerVolumeSpecName "kube-api-access-cb88l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:30 crc kubenswrapper[4972]: I0228 10:39:30.938286 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" (UID: "9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034817 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034888 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb88l\" (UniqueName: \"kubernetes.io/projected/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-kube-api-access-cb88l\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034911 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km5xh\" (UniqueName: \"kubernetes.io/projected/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-kube-api-access-km5xh\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034929 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034946 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034964 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034981 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.034997 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.035015 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.608326 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" event={"ID":"c728f38a-b1ad-4ce3-a3c4-a4a82596e70d","Type":"ContainerDied","Data":"c62f349dd8480e91ad7fa94eb67665c16381e705ee77c43cc0b23ffbbfe83d5a"} Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.608407 4972 scope.go:117] "RemoveContainer" containerID="8bfd915be868a8ace030a1d58dafe37d8c36dd2c77b375c6cea626a6247c6ef1" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.608593 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b95795899-cvf2j" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.612057 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" event={"ID":"9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4","Type":"ContainerDied","Data":"81bdb0e2474028fd03939625a49550d12278c15487f93d7bd77e72d0b5fb7431"} Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.612151 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.637136 4972 scope.go:117] "RemoveContainer" containerID="33cd4d8e80b487c5110ba2d646255f3def605f663411ee2abce815574f19ebb8" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.662718 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.679294 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66bdc57965-fxhd5"] Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.683819 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.687759 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5b95795899-cvf2j"] Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.795556 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" path="/var/lib/kubelet/pods/9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4/volumes" Feb 28 10:39:31 crc kubenswrapper[4972]: I0228 10:39:31.796524 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" path="/var/lib/kubelet/pods/c728f38a-b1ad-4ce3-a3c4-a4a82596e70d/volumes" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.085267 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.085895 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="extract-content" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.085929 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="extract-content" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.085959 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="extract-utilities" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.085980 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="extract-utilities" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086008 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" containerName="controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086026 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" containerName="controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086048 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086134 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086163 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" containerName="route-controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086180 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" containerName="route-controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086223 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" containerName="oc" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086240 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" containerName="oc" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086266 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086284 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086310 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="extract-content" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086326 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="extract-content" Feb 28 10:39:32 crc kubenswrapper[4972]: E0228 10:39:32.086554 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="extract-utilities" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086594 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="extract-utilities" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086842 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e05a811-1a21-48f9-899c-e1f172328e48" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086883 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bade0e0-1dde-4239-9afd-a4cfc285eecb" containerName="registry-server" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086909 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" containerName="oc" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086932 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c728f38a-b1ad-4ce3-a3c4-a4a82596e70d" containerName="controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.086965 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecf0af8-b7a9-4489-b848-51ab6dfcc1b4" containerName="route-controller-manager" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.087706 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.088973 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.090127 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.093815 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.093829 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.097380 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.097515 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.098149 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.098652 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.099430 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.099430 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.099801 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.100087 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.100789 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.104519 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.113961 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.114287 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.126038 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.149679 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.149773 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.149817 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk9mn\" (UniqueName: \"kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.149853 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.149952 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.150029 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.150078 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.150103 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvs77\" (UniqueName: \"kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.150139 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251121 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251191 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251222 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk9mn\" (UniqueName: \"kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251255 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.251955 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.252086 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.252441 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvs77\" (UniqueName: \"kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.252578 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.252640 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.254092 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.256651 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.256832 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.261425 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.261911 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.263334 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.281329 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvs77\" (UniqueName: \"kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77\") pod \"route-controller-manager-654cfc77c8-97q6n\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.281897 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk9mn\" (UniqueName: \"kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn\") pod \"controller-manager-fcf77db7d-gp6zh\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.423125 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.433337 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.885360 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:39:32 crc kubenswrapper[4972]: I0228 10:39:32.925182 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:39:32 crc kubenswrapper[4972]: W0228 10:39:32.936555 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47bdc14c_8ffa_46f8_833d_ffbcaa772e6d.slice/crio-606ce87b6f8b3a7eab52e7c4cd3147a9ddb0763493a83f0ac37463aba6cbccc9 WatchSource:0}: Error finding container 606ce87b6f8b3a7eab52e7c4cd3147a9ddb0763493a83f0ac37463aba6cbccc9: Status 404 returned error can't find the container with id 606ce87b6f8b3a7eab52e7c4cd3147a9ddb0763493a83f0ac37463aba6cbccc9 Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.241399 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.297013 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.627974 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" event={"ID":"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d","Type":"ContainerStarted","Data":"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228"} Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.628017 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" event={"ID":"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d","Type":"ContainerStarted","Data":"606ce87b6f8b3a7eab52e7c4cd3147a9ddb0763493a83f0ac37463aba6cbccc9"} Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.628320 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.629779 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" event={"ID":"d4174c65-1c1b-4d97-86f6-525749e1923f","Type":"ContainerStarted","Data":"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5"} Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.629838 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" event={"ID":"d4174c65-1c1b-4d97-86f6-525749e1923f","Type":"ContainerStarted","Data":"71fb14d1a5b8bcaca2f30bc74005f7ee6f76cb76482f04ded067d1fd4a1f8966"} Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.629915 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-85647" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="registry-server" containerID="cri-o://4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5" gracePeriod=2 Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.630027 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.635309 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.642935 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:39:33 crc kubenswrapper[4972]: I0228 10:39:33.646519 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" podStartSLOduration=3.646470308 podStartE2EDuration="3.646470308s" podCreationTimestamp="2026-02-28 10:39:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:39:33.644588395 +0000 UTC m=+230.556596133" watchObservedRunningTime="2026-02-28 10:39:33.646470308 +0000 UTC m=+230.558478046" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.006421 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.028065 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" podStartSLOduration=4.027991733 podStartE2EDuration="4.027991733s" podCreationTimestamp="2026-02-28 10:39:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:39:33.739680006 +0000 UTC m=+230.651687744" watchObservedRunningTime="2026-02-28 10:39:34.027991733 +0000 UTC m=+230.939999461" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.082388 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities\") pod \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.082499 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgtk4\" (UniqueName: \"kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4\") pod \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.082521 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content\") pod \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\" (UID: \"5cea766f-d665-474b-a7fa-5e193ce3c1fd\") " Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.083396 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities" (OuterVolumeSpecName: "utilities") pod "5cea766f-d665-474b-a7fa-5e193ce3c1fd" (UID: "5cea766f-d665-474b-a7fa-5e193ce3c1fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.087261 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4" (OuterVolumeSpecName: "kube-api-access-hgtk4") pod "5cea766f-d665-474b-a7fa-5e193ce3c1fd" (UID: "5cea766f-d665-474b-a7fa-5e193ce3c1fd"). InnerVolumeSpecName "kube-api-access-hgtk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.138689 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cea766f-d665-474b-a7fa-5e193ce3c1fd" (UID: "5cea766f-d665-474b-a7fa-5e193ce3c1fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.184236 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgtk4\" (UniqueName: \"kubernetes.io/projected/5cea766f-d665-474b-a7fa-5e193ce3c1fd-kube-api-access-hgtk4\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.184273 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.184304 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cea766f-d665-474b-a7fa-5e193ce3c1fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.420447 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9f545"] Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.638580 4972 generic.go:334] "Generic (PLEG): container finished" podID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerID="4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5" exitCode=0 Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.638693 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerDied","Data":"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5"} Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.639041 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85647" event={"ID":"5cea766f-d665-474b-a7fa-5e193ce3c1fd","Type":"ContainerDied","Data":"19fa698ae570b087cea9567de413259c2a108e2dc164e642a7701e366fca5b2f"} Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.639084 4972 scope.go:117] "RemoveContainer" containerID="4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.638769 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85647" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.662545 4972 scope.go:117] "RemoveContainer" containerID="c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.670766 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.674086 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-85647"] Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.686428 4972 scope.go:117] "RemoveContainer" containerID="11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.702718 4972 scope.go:117] "RemoveContainer" containerID="4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5" Feb 28 10:39:34 crc kubenswrapper[4972]: E0228 10:39:34.703292 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5\": container with ID starting with 4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5 not found: ID does not exist" containerID="4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.703326 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5"} err="failed to get container status \"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5\": rpc error: code = NotFound desc = could not find container \"4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5\": container with ID starting with 4532e75757f28993e274b0c0b80cf65b879df4d06599b354536ba80e1f46dea5 not found: ID does not exist" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.703348 4972 scope.go:117] "RemoveContainer" containerID="c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355" Feb 28 10:39:34 crc kubenswrapper[4972]: E0228 10:39:34.704867 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355\": container with ID starting with c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355 not found: ID does not exist" containerID="c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.704951 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355"} err="failed to get container status \"c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355\": rpc error: code = NotFound desc = could not find container \"c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355\": container with ID starting with c7452170520bb7c7e3c2c7c93fe2062822cd5de78b827af9340c63890e959355 not found: ID does not exist" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.705029 4972 scope.go:117] "RemoveContainer" containerID="11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348" Feb 28 10:39:34 crc kubenswrapper[4972]: E0228 10:39:34.705560 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348\": container with ID starting with 11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348 not found: ID does not exist" containerID="11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.705595 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348"} err="failed to get container status \"11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348\": rpc error: code = NotFound desc = could not find container \"11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348\": container with ID starting with 11f2a5fd53506061c44bd01a99ea793cdb919bd4bc110997bf94eb001476c348 not found: ID does not exist" Feb 28 10:39:34 crc kubenswrapper[4972]: I0228 10:39:34.848059 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.125005 4972 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.125348 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="extract-utilities" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.125368 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="extract-utilities" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.125438 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="registry-server" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.125454 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="registry-server" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.125498 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="extract-content" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.125511 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="extract-content" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.125697 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" containerName="registry-server" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126200 4972 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126535 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126709 4972 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.126931 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126942 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.126955 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126969 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.126978 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.126987 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127057 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127065 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127072 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127088 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127097 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127102 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127113 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127119 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127134 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127140 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127148 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127155 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127238 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8" gracePeriod=15 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127333 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127349 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127356 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127364 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127370 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127378 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127389 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127417 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54" gracePeriod=15 Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.127497 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127504 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127527 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178" gracePeriod=15 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127605 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127596 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9" gracePeriod=15 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127521 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957" gracePeriod=15 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.127816 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.136836 4972 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.178378 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197349 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197397 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197480 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197509 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197532 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197560 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197577 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.197728 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.223101 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299443 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299521 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299610 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299641 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299675 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299699 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299821 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299865 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299896 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299927 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299955 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.299983 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.300012 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.300041 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.461887 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:39:35 crc kubenswrapper[4972]: W0228 10:39:35.490349 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-834fc454afd4c06a5cc4d97d32550d6d952abc89047710dd3269650f4bd7b845 WatchSource:0}: Error finding container 834fc454afd4c06a5cc4d97d32550d6d952abc89047710dd3269650f4bd7b845: Status 404 returned error can't find the container with id 834fc454afd4c06a5cc4d97d32550d6d952abc89047710dd3269650f4bd7b845 Feb 28 10:39:35 crc kubenswrapper[4972]: E0228 10:39:35.496383 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189862e1d798596a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:39:35.494539626 +0000 UTC m=+232.406547364,LastTimestamp:2026-02-28 10:39:35.494539626 +0000 UTC m=+232.406547364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.649625 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.651348 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.652298 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54" exitCode=0 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.652342 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9" exitCode=0 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.652355 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957" exitCode=0 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.652366 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178" exitCode=2 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.652442 4972 scope.go:117] "RemoveContainer" containerID="6ca1eeb54a0e61cff3c52f9fa560f7fe6c7efff8b27bf41d60789cc0007dddd5" Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.655662 4972 generic.go:334] "Generic (PLEG): container finished" podID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" containerID="e695cbcf8a37b860dc918b4cea879365e51f1ebe111a288da4ef7c32e4b97281" exitCode=0 Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.655731 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d26eeeef-20d2-43c9-b132-9699eb84f9a4","Type":"ContainerDied","Data":"e695cbcf8a37b860dc918b4cea879365e51f1ebe111a288da4ef7c32e4b97281"} Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.657503 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"834fc454afd4c06a5cc4d97d32550d6d952abc89047710dd3269650f4bd7b845"} Feb 28 10:39:35 crc kubenswrapper[4972]: I0228 10:39:35.797646 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cea766f-d665-474b-a7fa-5e193ce3c1fd" path="/var/lib/kubelet/pods/5cea766f-d665-474b-a7fa-5e193ce3c1fd/volumes" Feb 28 10:39:36 crc kubenswrapper[4972]: I0228 10:39:36.672200 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b83834449beedcd962cb3965a74222de1932c944dccd1354368971a70b115fc6"} Feb 28 10:39:36 crc kubenswrapper[4972]: I0228 10:39:36.677827 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.110244 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.233113 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access\") pod \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.233216 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock\") pod \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.233361 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir\") pod \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\" (UID: \"d26eeeef-20d2-43c9-b132-9699eb84f9a4\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.233605 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock" (OuterVolumeSpecName: "var-lock") pod "d26eeeef-20d2-43c9-b132-9699eb84f9a4" (UID: "d26eeeef-20d2-43c9-b132-9699eb84f9a4"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.233673 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d26eeeef-20d2-43c9-b132-9699eb84f9a4" (UID: "d26eeeef-20d2-43c9-b132-9699eb84f9a4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.234205 4972 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-var-lock\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.234270 4972 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.254828 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d26eeeef-20d2-43c9-b132-9699eb84f9a4" (UID: "d26eeeef-20d2-43c9-b132-9699eb84f9a4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.335106 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d26eeeef-20d2-43c9-b132-9699eb84f9a4-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.588870 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.589820 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638183 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638260 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638298 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638305 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638316 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638482 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638885 4972 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638907 4972 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.638918 4972 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.696712 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.696729 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d26eeeef-20d2-43c9-b132-9699eb84f9a4","Type":"ContainerDied","Data":"7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7"} Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.696787 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e26665730867b24b6476ac02330e0ea21d66a761268c543f8114360ee8935d7" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.700645 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.701677 4972 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8" exitCode=0 Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.701783 4972 scope.go:117] "RemoveContainer" containerID="3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.701822 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.727929 4972 scope.go:117] "RemoveContainer" containerID="477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.757222 4972 scope.go:117] "RemoveContainer" containerID="99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.783438 4972 scope.go:117] "RemoveContainer" containerID="68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.802648 4972 scope.go:117] "RemoveContainer" containerID="85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.803366 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.822518 4972 scope.go:117] "RemoveContainer" containerID="db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.856014 4972 scope.go:117] "RemoveContainer" containerID="3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.856675 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54\": container with ID starting with 3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54 not found: ID does not exist" containerID="3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.856778 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54"} err="failed to get container status \"3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54\": rpc error: code = NotFound desc = could not find container \"3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54\": container with ID starting with 3808578930bb92da977cb22021f16bae0e167a9b79e118fe55182c44665f5f54 not found: ID does not exist" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.856842 4972 scope.go:117] "RemoveContainer" containerID="477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.857606 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9\": container with ID starting with 477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9 not found: ID does not exist" containerID="477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.857675 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9"} err="failed to get container status \"477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9\": rpc error: code = NotFound desc = could not find container \"477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9\": container with ID starting with 477c31b5136a1f48f3bf7c4029cc57014aa971f446d8a9c63c218c0f72435ee9 not found: ID does not exist" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.857716 4972 scope.go:117] "RemoveContainer" containerID="99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.858678 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957\": container with ID starting with 99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957 not found: ID does not exist" containerID="99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.858742 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957"} err="failed to get container status \"99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957\": rpc error: code = NotFound desc = could not find container \"99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957\": container with ID starting with 99e89924303627a4e214c8a76d37555eb82620489d1058543693264225ad2957 not found: ID does not exist" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.858774 4972 scope.go:117] "RemoveContainer" containerID="68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.859582 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178\": container with ID starting with 68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178 not found: ID does not exist" containerID="68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.859637 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178"} err="failed to get container status \"68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178\": rpc error: code = NotFound desc = could not find container \"68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178\": container with ID starting with 68067191e1fb4e1e15479188355be20500594547815981e2ca6af6dfea458178 not found: ID does not exist" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.859675 4972 scope.go:117] "RemoveContainer" containerID="85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.860062 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8\": container with ID starting with 85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8 not found: ID does not exist" containerID="85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.860111 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8"} err="failed to get container status \"85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8\": rpc error: code = NotFound desc = could not find container \"85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8\": container with ID starting with 85489b08fafdfb7b5388166918fd640f612431746ce24deade924ba733576bf8 not found: ID does not exist" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.860142 4972 scope.go:117] "RemoveContainer" containerID="db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747" Feb 28 10:39:37 crc kubenswrapper[4972]: E0228 10:39:37.860916 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747\": container with ID starting with db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747 not found: ID does not exist" containerID="db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747" Feb 28 10:39:37 crc kubenswrapper[4972]: I0228 10:39:37.860946 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747"} err="failed to get container status \"db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747\": rpc error: code = NotFound desc = could not find container \"db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747\": container with ID starting with db6f62e9b4b5ee287fc65f438e102bc7b3a1905a5a9095454a57ea4a0c325747 not found: ID does not exist" Feb 28 10:39:39 crc kubenswrapper[4972]: E0228 10:39:39.063575 4972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189862e1d798596a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-28 10:39:35.494539626 +0000 UTC m=+232.406547364,LastTimestamp:2026-02-28 10:39:35.494539626 +0000 UTC m=+232.406547364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 28 10:39:40 crc kubenswrapper[4972]: I0228 10:39:40.226320 4972 status_manager.go:851] "Failed to get status for pod" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" pod="openshift-marketplace/redhat-marketplace-bccmr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-bccmr\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:40 crc kubenswrapper[4972]: I0228 10:39:40.227169 4972 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:40 crc kubenswrapper[4972]: I0228 10:39:40.229042 4972 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:40 crc kubenswrapper[4972]: I0228 10:39:40.229579 4972 status_manager.go:851] "Failed to get status for pod" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:40 crc kubenswrapper[4972]: I0228 10:39:40.230094 4972 status_manager.go:851] "Failed to get status for pod" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" pod="openshift-marketplace/redhat-marketplace-bccmr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-bccmr\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:43 crc kubenswrapper[4972]: I0228 10:39:43.793075 4972 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:43 crc kubenswrapper[4972]: I0228 10:39:43.793693 4972 status_manager.go:851] "Failed to get status for pod" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:43 crc kubenswrapper[4972]: I0228 10:39:43.794061 4972 status_manager.go:851] "Failed to get status for pod" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" pod="openshift-marketplace/redhat-marketplace-bccmr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-bccmr\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.016619 4972 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.017688 4972 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.018211 4972 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.018653 4972 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.019121 4972 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:44 crc kubenswrapper[4972]: I0228 10:39:44.019168 4972 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.019433 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.220899 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Feb 28 10:39:44 crc kubenswrapper[4972]: E0228 10:39:44.622230 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Feb 28 10:39:45 crc kubenswrapper[4972]: E0228 10:39:45.423362 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.788784 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.792487 4972 status_manager.go:851] "Failed to get status for pod" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.794333 4972 status_manager.go:851] "Failed to get status for pod" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" pod="openshift-marketplace/redhat-marketplace-bccmr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-bccmr\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.795097 4972 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.821349 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.821413 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:45 crc kubenswrapper[4972]: E0228 10:39:45.821968 4972 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:45 crc kubenswrapper[4972]: I0228 10:39:45.822749 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:45 crc kubenswrapper[4972]: W0228 10:39:45.864078 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-38a2b07bf63b550b02b413b656a6c1024c0ffe7dacef52518d008d751baf1a96 WatchSource:0}: Error finding container 38a2b07bf63b550b02b413b656a6c1024c0ffe7dacef52518d008d751baf1a96: Status 404 returned error can't find the container with id 38a2b07bf63b550b02b413b656a6c1024c0ffe7dacef52518d008d751baf1a96 Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.789128 4972 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f0098766231b8490cc5d1b0382d42c9095632268fbf292839ff0d7667edd9b74" exitCode=0 Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.789279 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f0098766231b8490cc5d1b0382d42c9095632268fbf292839ff0d7667edd9b74"} Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.789962 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"38a2b07bf63b550b02b413b656a6c1024c0ffe7dacef52518d008d751baf1a96"} Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.790301 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.790316 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:46 crc kubenswrapper[4972]: E0228 10:39:46.790926 4972 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.791281 4972 status_manager.go:851] "Failed to get status for pod" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" pod="openshift-marketplace/redhat-marketplace-bccmr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-bccmr\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.791920 4972 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:46 crc kubenswrapper[4972]: I0228 10:39:46.792567 4972 status_manager.go:851] "Failed to get status for pod" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.12:6443: connect: connection refused" Feb 28 10:39:47 crc kubenswrapper[4972]: E0228 10:39:47.024473 4972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Feb 28 10:39:47 crc kubenswrapper[4972]: I0228 10:39:47.797639 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"03700b7350435150c4734eea7409c08b5336a2d208172b443e3e07c20c484032"} Feb 28 10:39:47 crc kubenswrapper[4972]: I0228 10:39:47.798079 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c89db9ad836a8f74401b79753501f87ee54fd4681040c88f61670f38a3b65590"} Feb 28 10:39:47 crc kubenswrapper[4972]: I0228 10:39:47.798090 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"222ff7f84bc5c1d5852434acc1e0265048afa4fbc1e375a3312899c31c4934c4"} Feb 28 10:39:48 crc kubenswrapper[4972]: I0228 10:39:48.812229 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"55138b4c9f6cf08286b0416f629052772bf5d76a109a952223a7ede423a403fa"} Feb 28 10:39:48 crc kubenswrapper[4972]: I0228 10:39:48.813126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4d1b3bb0c2b0821cfa50c6961f2313ce4dff49caa987a829ab1ae137ab95c202"} Feb 28 10:39:48 crc kubenswrapper[4972]: I0228 10:39:48.813631 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:48 crc kubenswrapper[4972]: I0228 10:39:48.813653 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:48 crc kubenswrapper[4972]: I0228 10:39:48.813990 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:49 crc kubenswrapper[4972]: I0228 10:39:49.818804 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Feb 28 10:39:49 crc kubenswrapper[4972]: I0228 10:39:49.820756 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 28 10:39:49 crc kubenswrapper[4972]: I0228 10:39:49.820807 4972 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="fc7cdead7cb07221486535a183c597d90a41bcc82bef09434fc5ac1ed2f0ddfb" exitCode=1 Feb 28 10:39:49 crc kubenswrapper[4972]: I0228 10:39:49.820844 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"fc7cdead7cb07221486535a183c597d90a41bcc82bef09434fc5ac1ed2f0ddfb"} Feb 28 10:39:49 crc kubenswrapper[4972]: I0228 10:39:49.821371 4972 scope.go:117] "RemoveContainer" containerID="fc7cdead7cb07221486535a183c597d90a41bcc82bef09434fc5ac1ed2f0ddfb" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.336181 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.823645 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.824063 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.830182 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.831632 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.833727 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 28 10:39:50 crc kubenswrapper[4972]: I0228 10:39:50.833800 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"375799866afabd4fc59e3ca5e6f48ff5a1194b57226328d70fbfb59f5c197884"} Feb 28 10:39:53 crc kubenswrapper[4972]: I0228 10:39:53.822507 4972 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:53 crc kubenswrapper[4972]: I0228 10:39:53.851026 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:53 crc kubenswrapper[4972]: I0228 10:39:53.851059 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:53 crc kubenswrapper[4972]: I0228 10:39:53.855737 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:39:53 crc kubenswrapper[4972]: I0228 10:39:53.913912 4972 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a075d654-00fc-4fe2-bd0f-cd626421447e" Feb 28 10:39:54 crc kubenswrapper[4972]: I0228 10:39:54.857125 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:54 crc kubenswrapper[4972]: I0228 10:39:54.857173 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:39:54 crc kubenswrapper[4972]: I0228 10:39:54.861736 4972 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a075d654-00fc-4fe2-bd0f-cd626421447e" Feb 28 10:39:57 crc kubenswrapper[4972]: I0228 10:39:57.130110 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:39:58 crc kubenswrapper[4972]: I0228 10:39:58.890271 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:39:58 crc kubenswrapper[4972]: I0228 10:39:58.890744 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:39:59 crc kubenswrapper[4972]: I0228 10:39:59.470350 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" containerID="cri-o://c6ff45262550ae89793c07ef093f0db31732d7fe21982ace0ab045397591cab1" gracePeriod=15 Feb 28 10:39:59 crc kubenswrapper[4972]: I0228 10:39:59.891503 4972 generic.go:334] "Generic (PLEG): container finished" podID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerID="c6ff45262550ae89793c07ef093f0db31732d7fe21982ace0ab045397591cab1" exitCode=0 Feb 28 10:39:59 crc kubenswrapper[4972]: I0228 10:39:59.891713 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" event={"ID":"63f17722-9aca-469f-85f4-3cb6774a90ea","Type":"ContainerDied","Data":"c6ff45262550ae89793c07ef093f0db31732d7fe21982ace0ab045397591cab1"} Feb 28 10:39:59 crc kubenswrapper[4972]: I0228 10:39:59.995595 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.063668 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.063760 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.063801 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.063836 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.064967 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.064993 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065076 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065235 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065299 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xshqs\" (UniqueName: \"kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065350 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065429 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065479 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065505 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065551 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065578 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065606 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065628 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca\") pod \"63f17722-9aca-469f-85f4-3cb6774a90ea\" (UID: \"63f17722-9aca-469f-85f4-3cb6774a90ea\") " Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065961 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065980 4972 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.065993 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.066390 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.067778 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.068521 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.072176 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.073226 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.074548 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.075515 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.075611 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs" (OuterVolumeSpecName: "kube-api-access-xshqs") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "kube-api-access-xshqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.078103 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.079401 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.079890 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.080221 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "63f17722-9aca-469f-85f4-3cb6774a90ea" (UID: "63f17722-9aca-469f-85f4-3cb6774a90ea"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.108625 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.167929 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168376 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168497 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xshqs\" (UniqueName: \"kubernetes.io/projected/63f17722-9aca-469f-85f4-3cb6774a90ea-kube-api-access-xshqs\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168587 4972 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168660 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168728 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168797 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168863 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168924 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.168988 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.169063 4972 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63f17722-9aca-469f-85f4-3cb6774a90ea-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.336307 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.340402 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.362543 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.366385 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.687857 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.898267 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.898275 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9f545" event={"ID":"63f17722-9aca-469f-85f4-3cb6774a90ea","Type":"ContainerDied","Data":"1df56dd5d529843394c853ca8041f9d9b7c9ca6c2d165d5761829ca551133984"} Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.898363 4972 scope.go:117] "RemoveContainer" containerID="c6ff45262550ae89793c07ef093f0db31732d7fe21982ace0ab045397591cab1" Feb 28 10:40:00 crc kubenswrapper[4972]: I0228 10:40:00.903763 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 28 10:40:01 crc kubenswrapper[4972]: I0228 10:40:01.107520 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 28 10:40:01 crc kubenswrapper[4972]: I0228 10:40:01.662393 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 28 10:40:01 crc kubenswrapper[4972]: I0228 10:40:01.680742 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 28 10:40:02 crc kubenswrapper[4972]: I0228 10:40:02.194532 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 28 10:40:02 crc kubenswrapper[4972]: I0228 10:40:02.712972 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 28 10:40:02 crc kubenswrapper[4972]: I0228 10:40:02.809699 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:02 crc kubenswrapper[4972]: I0228 10:40:02.847988 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 28 10:40:03 crc kubenswrapper[4972]: I0228 10:40:03.421669 4972 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 28 10:40:03 crc kubenswrapper[4972]: I0228 10:40:03.860567 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 28 10:40:04 crc kubenswrapper[4972]: I0228 10:40:04.244174 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 28 10:40:04 crc kubenswrapper[4972]: I0228 10:40:04.754510 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.465844 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.471518 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.563265 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.591766 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.719856 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.876294 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.921202 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 28 10:40:05 crc kubenswrapper[4972]: I0228 10:40:05.921209 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.064744 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.133815 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.592327 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.745626 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.805085 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.923809 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 28 10:40:06 crc kubenswrapper[4972]: I0228 10:40:06.972981 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 28 10:40:07 crc kubenswrapper[4972]: I0228 10:40:07.121177 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 28 10:40:07 crc kubenswrapper[4972]: I0228 10:40:07.357275 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 28 10:40:07 crc kubenswrapper[4972]: I0228 10:40:07.483514 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 28 10:40:07 crc kubenswrapper[4972]: I0228 10:40:07.785738 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 28 10:40:07 crc kubenswrapper[4972]: I0228 10:40:07.844206 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.112350 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.132818 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.218285 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.631025 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.648482 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.675383 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.778721 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.808328 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.845402 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.846350 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.865521 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.918452 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 28 10:40:08 crc kubenswrapper[4972]: I0228 10:40:08.950655 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.040572 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.066543 4972 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.104196 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.105874 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.222730 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.293002 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.336602 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.361656 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.533529 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.656335 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.675258 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.682093 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.722737 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.742611 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.755023 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.932957 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 28 10:40:09 crc kubenswrapper[4972]: I0228 10:40:09.958435 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.041501 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.061779 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.114850 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.205020 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.221276 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.258086 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.368767 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.390767 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.397083 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.438906 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.477741 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.589619 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.604071 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.698765 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.738235 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.778017 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.828105 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.903443 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.935770 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 28 10:40:10 crc kubenswrapper[4972]: I0228 10:40:10.951278 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.045735 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.100833 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.119728 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.266746 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.317151 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.461140 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.604744 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.613770 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.628612 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.637103 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.682770 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.757428 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.825451 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.869388 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.904050 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.910546 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 28 10:40:11 crc kubenswrapper[4972]: I0228 10:40:11.990570 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.032333 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.107147 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.386591 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.424557 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.437826 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.478599 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.625652 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.715986 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.756605 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.773064 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.811657 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.849911 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.860340 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.864651 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.916529 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 28 10:40:12 crc kubenswrapper[4972]: I0228 10:40:12.925795 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.051066 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.152531 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.152533 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.230316 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.246252 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.291649 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.313994 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.395287 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.475138 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.488671 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.529663 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.562773 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.697117 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.759194 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.889449 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.920626 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.980568 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 28 10:40:13 crc kubenswrapper[4972]: I0228 10:40:13.999863 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.040620 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.158381 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.158637 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.178990 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.198707 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.231433 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.241270 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.252321 4972 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.294110 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.305007 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.391894 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.422932 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.422971 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.462083 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.524729 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.603779 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.768824 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.800405 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.813721 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 28 10:40:14 crc kubenswrapper[4972]: I0228 10:40:14.841327 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.059241 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.075297 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.120866 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.195379 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.286036 4972 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.705569 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.751036 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.752016 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.779611 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.804496 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 28 10:40:15 crc kubenswrapper[4972]: I0228 10:40:15.849284 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.065187 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.118599 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.204119 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.212583 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.365702 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.501341 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.516824 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.526368 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.600338 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.719158 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.727484 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.774272 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.828353 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.851791 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.853897 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 28 10:40:16 crc kubenswrapper[4972]: I0228 10:40:16.982414 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.036615 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.058481 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.069981 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.071440 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.085665 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.244522 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.268810 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.280008 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.641370 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.683776 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.699439 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.785192 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.880921 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.974572 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 28 10:40:17 crc kubenswrapper[4972]: I0228 10:40:17.978310 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.266991 4972 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.269257 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.269227664 podStartE2EDuration="43.269227664s" podCreationTimestamp="2026-02-28 10:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:39:53.8622312 +0000 UTC m=+250.774238938" watchObservedRunningTime="2026-02-28 10:40:18.269227664 +0000 UTC m=+275.181235403" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273029 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9f545","openshift-kube-apiserver/kube-apiserver-crc"] Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273103 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-infra/auto-csr-approver-29537920-mpsj6","openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f"] Feb 28 10:40:18 crc kubenswrapper[4972]: E0228 10:40:18.273356 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273372 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" Feb 28 10:40:18 crc kubenswrapper[4972]: E0228 10:40:18.273398 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" containerName="installer" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273406 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" containerName="installer" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273701 4972 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.273748 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="49012204-eddc-40a8-a930-c82043f8e851" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.274029 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" containerName="oauth-openshift" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.274104 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d26eeeef-20d2-43c9-b132-9699eb84f9a4" containerName="installer" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.276313 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.277249 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.279011 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.284430 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.284767 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.284815 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.285403 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.285876 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.285918 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.286006 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.286985 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.291491 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.291705 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.291507 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.292010 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.292018 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.292101 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.292192 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.299182 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.303442 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.304038 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.316662 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.327317 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.327284101 podStartE2EDuration="25.327284101s" podCreationTimestamp="2026-02-28 10:39:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:18.323050674 +0000 UTC m=+275.235058452" watchObservedRunningTime="2026-02-28 10:40:18.327284101 +0000 UTC m=+275.239291869" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384700 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-error\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384802 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384842 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384905 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-login\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384932 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384962 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.384983 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385008 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-session\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385035 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385068 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-policies\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385106 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-dir\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385134 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chgz9\" (UniqueName: \"kubernetes.io/projected/44ab3663-512a-47bd-8e55-1ef35dd32a4d-kube-api-access-chgz9\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385162 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385197 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj6kj\" (UniqueName: \"kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj\") pod \"auto-csr-approver-29537920-mpsj6\" (UID: \"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e\") " pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.385223 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.473648 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486572 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-dir\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486668 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chgz9\" (UniqueName: \"kubernetes.io/projected/44ab3663-512a-47bd-8e55-1ef35dd32a4d-kube-api-access-chgz9\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486724 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-dir\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486733 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486891 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj6kj\" (UniqueName: \"kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj\") pod \"auto-csr-approver-29537920-mpsj6\" (UID: \"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e\") " pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486937 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.486964 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-error\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487010 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487053 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487211 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-login\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487239 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487290 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487316 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487378 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-session\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487414 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.487513 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-policies\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.488222 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.488636 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-audit-policies\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.489283 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.489916 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.495396 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.496001 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-error\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.496092 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.496967 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.499226 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-login\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.499788 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.500073 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-system-session\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.502754 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/44ab3663-512a-47bd-8e55-1ef35dd32a4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.508356 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chgz9\" (UniqueName: \"kubernetes.io/projected/44ab3663-512a-47bd-8e55-1ef35dd32a4d-kube-api-access-chgz9\") pod \"oauth-openshift-b7d5b84cf-k5n2f\" (UID: \"44ab3663-512a-47bd-8e55-1ef35dd32a4d\") " pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.511945 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj6kj\" (UniqueName: \"kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj\") pod \"auto-csr-approver-29537920-mpsj6\" (UID: \"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e\") " pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.574225 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.618163 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.634094 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.687788 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.739451 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.786355 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.902354 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 28 10:40:18 crc kubenswrapper[4972]: I0228 10:40:18.904069 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.024937 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.114779 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537920-mpsj6"] Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.171611 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.171722 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f"] Feb 28 10:40:19 crc kubenswrapper[4972]: W0228 10:40:19.173376 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44ab3663_512a_47bd_8e55_1ef35dd32a4d.slice/crio-4c49803e28f8f16fd4658aa88e0130680d18981c67d4323fb63422a6806fc4bc WatchSource:0}: Error finding container 4c49803e28f8f16fd4658aa88e0130680d18981c67d4323fb63422a6806fc4bc: Status 404 returned error can't find the container with id 4c49803e28f8f16fd4658aa88e0130680d18981c67d4323fb63422a6806fc4bc Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.231825 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.419566 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.458217 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.525675 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.554866 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.671928 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.725217 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.796851 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63f17722-9aca-469f-85f4-3cb6774a90ea" path="/var/lib/kubelet/pods/63f17722-9aca-469f-85f4-3cb6774a90ea/volumes" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.840915 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 28 10:40:19 crc kubenswrapper[4972]: I0228 10:40:19.884807 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.030373 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" event={"ID":"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e","Type":"ContainerStarted","Data":"f4d2144d35a324a6bf9d387ac536ec8856cac8c4d798faa26c8c99d9ece7a23b"} Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.036171 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-b7d5b84cf-k5n2f_44ab3663-512a-47bd-8e55-1ef35dd32a4d/oauth-openshift/0.log" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.036533 4972 generic.go:334] "Generic (PLEG): container finished" podID="44ab3663-512a-47bd-8e55-1ef35dd32a4d" containerID="2e12ca2cbc3e6419ddf3bd9737265e205b1d72d6fb5b060760b7837660171214" exitCode=255 Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.036585 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" event={"ID":"44ab3663-512a-47bd-8e55-1ef35dd32a4d","Type":"ContainerDied","Data":"2e12ca2cbc3e6419ddf3bd9737265e205b1d72d6fb5b060760b7837660171214"} Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.036616 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" event={"ID":"44ab3663-512a-47bd-8e55-1ef35dd32a4d","Type":"ContainerStarted","Data":"4c49803e28f8f16fd4658aa88e0130680d18981c67d4323fb63422a6806fc4bc"} Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.038826 4972 scope.go:117] "RemoveContainer" containerID="2e12ca2cbc3e6419ddf3bd9737265e205b1d72d6fb5b060760b7837660171214" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.187129 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.326356 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.415497 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.643877 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.657381 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 28 10:40:20 crc kubenswrapper[4972]: I0228 10:40:20.919685 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.035563 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.044606 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.048118 4972 generic.go:334] "Generic (PLEG): container finished" podID="4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" containerID="6a85130a8dd8f8b8faf3e7d1949e41c9bb84368575780ddb4d5df8bf0e86c615" exitCode=0 Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.048259 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" event={"ID":"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e","Type":"ContainerDied","Data":"6a85130a8dd8f8b8faf3e7d1949e41c9bb84368575780ddb4d5df8bf0e86c615"} Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.052682 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-b7d5b84cf-k5n2f_44ab3663-512a-47bd-8e55-1ef35dd32a4d/oauth-openshift/0.log" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.052747 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" event={"ID":"44ab3663-512a-47bd-8e55-1ef35dd32a4d","Type":"ContainerStarted","Data":"44e3e04bbd5989fe45ca22f4b6e1188a30fb528ff519e3acdcbb64d8a8caa66c"} Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.053210 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.110757 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" podStartSLOduration=47.110704201 podStartE2EDuration="47.110704201s" podCreationTimestamp="2026-02-28 10:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:21.107723821 +0000 UTC m=+278.019731599" watchObservedRunningTime="2026-02-28 10:40:21.110704201 +0000 UTC m=+278.022711969" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.169010 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.242172 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-b7d5b84cf-k5n2f" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.403051 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.636632 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 28 10:40:21 crc kubenswrapper[4972]: I0228 10:40:21.941906 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.250092 4972 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.457992 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.498975 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.547426 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj6kj\" (UniqueName: \"kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj\") pod \"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e\" (UID: \"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e\") " Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.558483 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj" (OuterVolumeSpecName: "kube-api-access-gj6kj") pod "4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" (UID: "4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e"). InnerVolumeSpecName "kube-api-access-gj6kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.649223 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj6kj\" (UniqueName: \"kubernetes.io/projected/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e-kube-api-access-gj6kj\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.903262 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.933302 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.957234 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 28 10:40:22 crc kubenswrapper[4972]: I0228 10:40:22.986923 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 28 10:40:23 crc kubenswrapper[4972]: I0228 10:40:23.071251 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" event={"ID":"4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e","Type":"ContainerDied","Data":"f4d2144d35a324a6bf9d387ac536ec8856cac8c4d798faa26c8c99d9ece7a23b"} Feb 28 10:40:23 crc kubenswrapper[4972]: I0228 10:40:23.071315 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537920-mpsj6" Feb 28 10:40:23 crc kubenswrapper[4972]: I0228 10:40:23.071331 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4d2144d35a324a6bf9d387ac536ec8856cac8c4d798faa26c8c99d9ece7a23b" Feb 28 10:40:27 crc kubenswrapper[4972]: I0228 10:40:27.605976 4972 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 28 10:40:27 crc kubenswrapper[4972]: I0228 10:40:27.606534 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b83834449beedcd962cb3965a74222de1932c944dccd1354368971a70b115fc6" gracePeriod=5 Feb 28 10:40:28 crc kubenswrapper[4972]: I0228 10:40:28.891194 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:40:28 crc kubenswrapper[4972]: I0228 10:40:28.891581 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.335013 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.336420 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" podUID="d4174c65-1c1b-4d97-86f6-525749e1923f" containerName="controller-manager" containerID="cri-o://2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5" gracePeriod=30 Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.435289 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.435538 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" podUID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" containerName="route-controller-manager" containerID="cri-o://1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228" gracePeriod=30 Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.795603 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.801543 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980114 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca\") pod \"d4174c65-1c1b-4d97-86f6-525749e1923f\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980226 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert\") pod \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980323 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk9mn\" (UniqueName: \"kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn\") pod \"d4174c65-1c1b-4d97-86f6-525749e1923f\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980355 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config\") pod \"d4174c65-1c1b-4d97-86f6-525749e1923f\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980390 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert\") pod \"d4174c65-1c1b-4d97-86f6-525749e1923f\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980418 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config\") pod \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980470 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvs77\" (UniqueName: \"kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77\") pod \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980501 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles\") pod \"d4174c65-1c1b-4d97-86f6-525749e1923f\" (UID: \"d4174c65-1c1b-4d97-86f6-525749e1923f\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.980528 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca\") pod \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\" (UID: \"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d\") " Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.981815 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca" (OuterVolumeSpecName: "client-ca") pod "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" (UID: "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.982332 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4174c65-1c1b-4d97-86f6-525749e1923f" (UID: "d4174c65-1c1b-4d97-86f6-525749e1923f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.984268 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d4174c65-1c1b-4d97-86f6-525749e1923f" (UID: "d4174c65-1c1b-4d97-86f6-525749e1923f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.984786 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config" (OuterVolumeSpecName: "config") pod "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" (UID: "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:30 crc kubenswrapper[4972]: I0228 10:40:30.984828 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config" (OuterVolumeSpecName: "config") pod "d4174c65-1c1b-4d97-86f6-525749e1923f" (UID: "d4174c65-1c1b-4d97-86f6-525749e1923f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.003807 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4174c65-1c1b-4d97-86f6-525749e1923f" (UID: "d4174c65-1c1b-4d97-86f6-525749e1923f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.003895 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77" (OuterVolumeSpecName: "kube-api-access-vvs77") pod "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" (UID: "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d"). InnerVolumeSpecName "kube-api-access-vvs77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.004000 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" (UID: "47bdc14c-8ffa-46f8-833d-ffbcaa772e6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.004964 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn" (OuterVolumeSpecName: "kube-api-access-vk9mn") pod "d4174c65-1c1b-4d97-86f6-525749e1923f" (UID: "d4174c65-1c1b-4d97-86f6-525749e1923f"). InnerVolumeSpecName "kube-api-access-vk9mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.081834 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk9mn\" (UniqueName: \"kubernetes.io/projected/d4174c65-1c1b-4d97-86f6-525749e1923f-kube-api-access-vk9mn\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082092 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082156 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4174c65-1c1b-4d97-86f6-525749e1923f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082216 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082296 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvs77\" (UniqueName: \"kubernetes.io/projected/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-kube-api-access-vvs77\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082353 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082406 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082493 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4174c65-1c1b-4d97-86f6-525749e1923f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.082580 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.125439 4972 generic.go:334] "Generic (PLEG): container finished" podID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" containerID="1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228" exitCode=0 Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.125567 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" event={"ID":"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d","Type":"ContainerDied","Data":"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228"} Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.125877 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" event={"ID":"47bdc14c-8ffa-46f8-833d-ffbcaa772e6d","Type":"ContainerDied","Data":"606ce87b6f8b3a7eab52e7c4cd3147a9ddb0763493a83f0ac37463aba6cbccc9"} Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.125615 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.125927 4972 scope.go:117] "RemoveContainer" containerID="1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.127399 4972 generic.go:334] "Generic (PLEG): container finished" podID="d4174c65-1c1b-4d97-86f6-525749e1923f" containerID="2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5" exitCode=0 Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.127449 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" event={"ID":"d4174c65-1c1b-4d97-86f6-525749e1923f","Type":"ContainerDied","Data":"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5"} Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.127510 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.127542 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-fcf77db7d-gp6zh" event={"ID":"d4174c65-1c1b-4d97-86f6-525749e1923f","Type":"ContainerDied","Data":"71fb14d1a5b8bcaca2f30bc74005f7ee6f76cb76482f04ded067d1fd4a1f8966"} Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.149721 4972 scope.go:117] "RemoveContainer" containerID="1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228" Feb 28 10:40:31 crc kubenswrapper[4972]: E0228 10:40:31.150145 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228\": container with ID starting with 1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228 not found: ID does not exist" containerID="1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.150183 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228"} err="failed to get container status \"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228\": rpc error: code = NotFound desc = could not find container \"1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228\": container with ID starting with 1f75059488e286edf7ee9a640a8e4e822e91d37ff2fe334e352f6e3798494228 not found: ID does not exist" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.150210 4972 scope.go:117] "RemoveContainer" containerID="2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.161242 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.172868 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-fcf77db7d-gp6zh"] Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.173295 4972 scope.go:117] "RemoveContainer" containerID="2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5" Feb 28 10:40:31 crc kubenswrapper[4972]: E0228 10:40:31.173788 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5\": container with ID starting with 2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5 not found: ID does not exist" containerID="2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.173835 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5"} err="failed to get container status \"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5\": rpc error: code = NotFound desc = could not find container \"2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5\": container with ID starting with 2a89efdb406d69c4e52018a637ea37c95755bcf5e62514659cfad8ee0d4150b5 not found: ID does not exist" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.176424 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.181641 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-654cfc77c8-97q6n"] Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.801583 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" path="/var/lib/kubelet/pods/47bdc14c-8ffa-46f8-833d-ffbcaa772e6d/volumes" Feb 28 10:40:31 crc kubenswrapper[4972]: I0228 10:40:31.802831 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4174c65-1c1b-4d97-86f6-525749e1923f" path="/var/lib/kubelet/pods/d4174c65-1c1b-4d97-86f6-525749e1923f/volumes" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136151 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:32 crc kubenswrapper[4972]: E0228 10:40:32.136535 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136557 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 28 10:40:32 crc kubenswrapper[4972]: E0228 10:40:32.136580 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" containerName="oc" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136618 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" containerName="oc" Feb 28 10:40:32 crc kubenswrapper[4972]: E0228 10:40:32.136648 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" containerName="route-controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136661 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" containerName="route-controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: E0228 10:40:32.136686 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4174c65-1c1b-4d97-86f6-525749e1923f" containerName="controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136698 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4174c65-1c1b-4d97-86f6-525749e1923f" containerName="controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136864 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136883 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" containerName="oc" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136900 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4174c65-1c1b-4d97-86f6-525749e1923f" containerName="controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.136918 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bdc14c-8ffa-46f8-833d-ffbcaa772e6d" containerName="route-controller-manager" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.137520 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.141773 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.142191 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.142410 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.142436 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.142956 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.143237 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.149567 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.150817 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.153925 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.154148 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.154220 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.154717 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.154759 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.157425 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.158517 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.162522 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.169385 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299301 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299413 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299454 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299594 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26hrq\" (UniqueName: \"kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299646 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299866 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.299989 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkgvl\" (UniqueName: \"kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.300039 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.300121 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401716 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401817 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkgvl\" (UniqueName: \"kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401911 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401946 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.401999 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.402029 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.402063 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.402107 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26hrq\" (UniqueName: \"kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.404126 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.404845 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.405821 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.406847 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.409009 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.416436 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.420380 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.433774 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkgvl\" (UniqueName: \"kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl\") pod \"route-controller-manager-745bb64d9b-z2r7p\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.441734 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26hrq\" (UniqueName: \"kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq\") pod \"controller-manager-f746d79b7-45hv4\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.459327 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.479739 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:32 crc kubenswrapper[4972]: I0228 10:40:32.737563 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:32 crc kubenswrapper[4972]: W0228 10:40:32.743731 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67808ae3_db0c_4498_8b7d_a1a51d9d486d.slice/crio-df5c81849adb5a8692430026ef1462a906d5b9cfcfef55c7ce13183a07c028ba WatchSource:0}: Error finding container df5c81849adb5a8692430026ef1462a906d5b9cfcfef55c7ce13183a07c028ba: Status 404 returned error can't find the container with id df5c81849adb5a8692430026ef1462a906d5b9cfcfef55c7ce13183a07c028ba Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.010583 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.154010 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.154055 4972 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b83834449beedcd962cb3965a74222de1932c944dccd1354368971a70b115fc6" exitCode=137 Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.156397 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" event={"ID":"64cd8d61-482e-46d2-8895-7efe28ddd6f5","Type":"ContainerStarted","Data":"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79"} Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.156625 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" event={"ID":"64cd8d61-482e-46d2-8895-7efe28ddd6f5","Type":"ContainerStarted","Data":"63dcf6ad0500a45677f32d50b0f482a57577c1005564eb7e24955442d6f9d739"} Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.157777 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" event={"ID":"67808ae3-db0c-4498-8b7d-a1a51d9d486d","Type":"ContainerStarted","Data":"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b"} Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.157807 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" event={"ID":"67808ae3-db0c-4498-8b7d-a1a51d9d486d","Type":"ContainerStarted","Data":"df5c81849adb5a8692430026ef1462a906d5b9cfcfef55c7ce13183a07c028ba"} Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.158147 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.160510 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.160599 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.174935 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" podStartSLOduration=3.174916192 podStartE2EDuration="3.174916192s" podCreationTimestamp="2026-02-28 10:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:33.172186919 +0000 UTC m=+290.084194677" watchObservedRunningTime="2026-02-28 10:40:33.174916192 +0000 UTC m=+290.086923920" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.317795 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.317851 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.317886 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.317954 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.318022 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.319102 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.319142 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.319168 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.319192 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.332312 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.419490 4972 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.419531 4972 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.419544 4972 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.419559 4972 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.419571 4972 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.547767 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.797883 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.798344 4972 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.810553 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.810594 4972 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="951c7063-19c5-490b-8671-81a05084aa77" Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.812652 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 28 10:40:33 crc kubenswrapper[4972]: I0228 10:40:33.812750 4972 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="951c7063-19c5-490b-8671-81a05084aa77" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.165500 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.166366 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.166768 4972 scope.go:117] "RemoveContainer" containerID="b83834449beedcd962cb3965a74222de1932c944dccd1354368971a70b115fc6" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.167000 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.180765 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:34 crc kubenswrapper[4972]: I0228 10:40:34.188017 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" podStartSLOduration=4.187990848 podStartE2EDuration="4.187990848s" podCreationTimestamp="2026-02-28 10:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:34.186186753 +0000 UTC m=+291.098194511" watchObservedRunningTime="2026-02-28 10:40:34.187990848 +0000 UTC m=+291.099998596" Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.300100 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.301198 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" podUID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" containerName="controller-manager" containerID="cri-o://3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79" gracePeriod=30 Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.334623 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.334907 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" podUID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" containerName="route-controller-manager" containerID="cri-o://c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b" gracePeriod=30 Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.901658 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:50 crc kubenswrapper[4972]: I0228 10:40:50.945902 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.068976 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca\") pod \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069035 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config\") pod \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069095 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26hrq\" (UniqueName: \"kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq\") pod \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069146 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles\") pod \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069178 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert\") pod \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069229 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert\") pod \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069275 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca\") pod \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069301 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config\") pod \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\" (UID: \"64cd8d61-482e-46d2-8895-7efe28ddd6f5\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.069346 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkgvl\" (UniqueName: \"kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl\") pod \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\" (UID: \"67808ae3-db0c-4498-8b7d-a1a51d9d486d\") " Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070051 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca" (OuterVolumeSpecName: "client-ca") pod "64cd8d61-482e-46d2-8895-7efe28ddd6f5" (UID: "64cd8d61-482e-46d2-8895-7efe28ddd6f5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070097 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config" (OuterVolumeSpecName: "config") pod "67808ae3-db0c-4498-8b7d-a1a51d9d486d" (UID: "67808ae3-db0c-4498-8b7d-a1a51d9d486d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070613 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca" (OuterVolumeSpecName: "client-ca") pod "67808ae3-db0c-4498-8b7d-a1a51d9d486d" (UID: "67808ae3-db0c-4498-8b7d-a1a51d9d486d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070698 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070720 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.070832 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "64cd8d61-482e-46d2-8895-7efe28ddd6f5" (UID: "64cd8d61-482e-46d2-8895-7efe28ddd6f5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.071419 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config" (OuterVolumeSpecName: "config") pod "64cd8d61-482e-46d2-8895-7efe28ddd6f5" (UID: "64cd8d61-482e-46d2-8895-7efe28ddd6f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.074614 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "64cd8d61-482e-46d2-8895-7efe28ddd6f5" (UID: "64cd8d61-482e-46d2-8895-7efe28ddd6f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.075252 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq" (OuterVolumeSpecName: "kube-api-access-26hrq") pod "64cd8d61-482e-46d2-8895-7efe28ddd6f5" (UID: "64cd8d61-482e-46d2-8895-7efe28ddd6f5"). InnerVolumeSpecName "kube-api-access-26hrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.075388 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "67808ae3-db0c-4498-8b7d-a1a51d9d486d" (UID: "67808ae3-db0c-4498-8b7d-a1a51d9d486d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.076011 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl" (OuterVolumeSpecName: "kube-api-access-zkgvl") pod "67808ae3-db0c-4498-8b7d-a1a51d9d486d" (UID: "67808ae3-db0c-4498-8b7d-a1a51d9d486d"). InnerVolumeSpecName "kube-api-access-zkgvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171876 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26hrq\" (UniqueName: \"kubernetes.io/projected/64cd8d61-482e-46d2-8895-7efe28ddd6f5-kube-api-access-26hrq\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171916 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171926 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cd8d61-482e-46d2-8895-7efe28ddd6f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171936 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67808ae3-db0c-4498-8b7d-a1a51d9d486d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171944 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67808ae3-db0c-4498-8b7d-a1a51d9d486d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171952 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cd8d61-482e-46d2-8895-7efe28ddd6f5-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.171960 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkgvl\" (UniqueName: \"kubernetes.io/projected/67808ae3-db0c-4498-8b7d-a1a51d9d486d-kube-api-access-zkgvl\") on node \"crc\" DevicePath \"\"" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.285588 4972 generic.go:334] "Generic (PLEG): container finished" podID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" containerID="3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79" exitCode=0 Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.285683 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.285705 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" event={"ID":"64cd8d61-482e-46d2-8895-7efe28ddd6f5","Type":"ContainerDied","Data":"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79"} Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.285783 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-45hv4" event={"ID":"64cd8d61-482e-46d2-8895-7efe28ddd6f5","Type":"ContainerDied","Data":"63dcf6ad0500a45677f32d50b0f482a57577c1005564eb7e24955442d6f9d739"} Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.285806 4972 scope.go:117] "RemoveContainer" containerID="3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.288126 4972 generic.go:334] "Generic (PLEG): container finished" podID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" containerID="c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b" exitCode=0 Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.288184 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.288168 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" event={"ID":"67808ae3-db0c-4498-8b7d-a1a51d9d486d","Type":"ContainerDied","Data":"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b"} Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.288323 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p" event={"ID":"67808ae3-db0c-4498-8b7d-a1a51d9d486d","Type":"ContainerDied","Data":"df5c81849adb5a8692430026ef1462a906d5b9cfcfef55c7ce13183a07c028ba"} Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.321007 4972 scope.go:117] "RemoveContainer" containerID="3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79" Feb 28 10:40:51 crc kubenswrapper[4972]: E0228 10:40:51.328954 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79\": container with ID starting with 3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79 not found: ID does not exist" containerID="3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.328995 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79"} err="failed to get container status \"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79\": rpc error: code = NotFound desc = could not find container \"3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79\": container with ID starting with 3d06619da14b2ef830f1eafe44acdb6e02eb0bc9e8a1433ec54a56135decbe79 not found: ID does not exist" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.329016 4972 scope.go:117] "RemoveContainer" containerID="c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.331949 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.338108 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-z2r7p"] Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.343273 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.347674 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-45hv4"] Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.349203 4972 scope.go:117] "RemoveContainer" containerID="c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b" Feb 28 10:40:51 crc kubenswrapper[4972]: E0228 10:40:51.349626 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b\": container with ID starting with c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b not found: ID does not exist" containerID="c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.349663 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b"} err="failed to get container status \"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b\": rpc error: code = NotFound desc = could not find container \"c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b\": container with ID starting with c75924b7036fe03bb43d904175475e1680026fd2119b3b7c976cf45418ba544b not found: ID does not exist" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.801039 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" path="/var/lib/kubelet/pods/64cd8d61-482e-46d2-8895-7efe28ddd6f5/volumes" Feb 28 10:40:51 crc kubenswrapper[4972]: I0228 10:40:51.801693 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" path="/var/lib/kubelet/pods/67808ae3-db0c-4498-8b7d-a1a51d9d486d/volumes" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.148829 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:40:52 crc kubenswrapper[4972]: E0228 10:40:52.149221 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" containerName="route-controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.149239 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" containerName="route-controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: E0228 10:40:52.149256 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" containerName="controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.149263 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" containerName="controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.149378 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="67808ae3-db0c-4498-8b7d-a1a51d9d486d" containerName="route-controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.149388 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="64cd8d61-482e-46d2-8895-7efe28ddd6f5" containerName="controller-manager" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.149980 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.155452 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.155611 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.155843 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.156031 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.156166 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.156280 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.157380 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.158448 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.161124 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.161485 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.162860 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.163006 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.163122 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.163517 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.164342 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.169526 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.172653 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291138 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291180 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291205 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291233 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291248 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291277 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291356 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j7ww\" (UniqueName: \"kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291383 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.291429 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t7vb\" (UniqueName: \"kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392748 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t7vb\" (UniqueName: \"kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392808 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392835 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392852 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392879 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392897 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392922 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392940 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j7ww\" (UniqueName: \"kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.392955 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.394032 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.394162 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.394613 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.395449 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.395666 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.401742 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.407625 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.425927 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j7ww\" (UniqueName: \"kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww\") pod \"controller-manager-759b4c7f9b-dstl6\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.428136 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t7vb\" (UniqueName: \"kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb\") pod \"route-controller-manager-5479c9f4cd-lxvd6\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.505380 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.519021 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:52 crc kubenswrapper[4972]: I0228 10:40:52.984244 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.046206 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:40:53 crc kubenswrapper[4972]: W0228 10:40:53.073413 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff2eada_b82e_46d9_afc3_8ab91f6ad064.slice/crio-5f02f5e7acfe7f7a78fbcece5684071b1be17a80642d518a8a932e695afcb9ab WatchSource:0}: Error finding container 5f02f5e7acfe7f7a78fbcece5684071b1be17a80642d518a8a932e695afcb9ab: Status 404 returned error can't find the container with id 5f02f5e7acfe7f7a78fbcece5684071b1be17a80642d518a8a932e695afcb9ab Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.312067 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" event={"ID":"a0ee3613-ba29-4242-8691-2badf68ec79a","Type":"ContainerStarted","Data":"74165d85c7fd2b02776e667c504ada86563513f9a73c1aee02cc712d62618b93"} Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.312157 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" event={"ID":"a0ee3613-ba29-4242-8691-2badf68ec79a","Type":"ContainerStarted","Data":"345fefd81d0ff86c4a6ed19556d153d80b91cc0477812c74d8ba149e859de7b2"} Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.312628 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.316028 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" event={"ID":"cff2eada-b82e-46d9-afc3-8ab91f6ad064","Type":"ContainerStarted","Data":"8f99513e9c152428d8113ed5e4e8f88c328b4f33fce0c3cf5c9ed97668738280"} Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.316062 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" event={"ID":"cff2eada-b82e-46d9-afc3-8ab91f6ad064","Type":"ContainerStarted","Data":"5f02f5e7acfe7f7a78fbcece5684071b1be17a80642d518a8a932e695afcb9ab"} Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.316283 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.326686 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.335506 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" podStartSLOduration=3.335485772 podStartE2EDuration="3.335485772s" podCreationTimestamp="2026-02-28 10:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:53.33341461 +0000 UTC m=+310.245422348" watchObservedRunningTime="2026-02-28 10:40:53.335485772 +0000 UTC m=+310.247493550" Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.360501 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" podStartSLOduration=3.3604800839999998 podStartE2EDuration="3.360480084s" podCreationTimestamp="2026-02-28 10:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:40:53.357081792 +0000 UTC m=+310.269089560" watchObservedRunningTime="2026-02-28 10:40:53.360480084 +0000 UTC m=+310.272487822" Feb 28 10:40:53 crc kubenswrapper[4972]: I0228 10:40:53.952398 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:40:58 crc kubenswrapper[4972]: I0228 10:40:58.890266 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:40:58 crc kubenswrapper[4972]: I0228 10:40:58.890787 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:40:58 crc kubenswrapper[4972]: I0228 10:40:58.890859 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:40:58 crc kubenswrapper[4972]: I0228 10:40:58.892011 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:40:58 crc kubenswrapper[4972]: I0228 10:40:58.892153 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094" gracePeriod=600 Feb 28 10:40:59 crc kubenswrapper[4972]: I0228 10:40:59.356103 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094" exitCode=0 Feb 28 10:40:59 crc kubenswrapper[4972]: I0228 10:40:59.356145 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094"} Feb 28 10:40:59 crc kubenswrapper[4972]: I0228 10:40:59.356177 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd"} Feb 28 10:41:38 crc kubenswrapper[4972]: I0228 10:41:38.337976 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:41:38 crc kubenswrapper[4972]: I0228 10:41:38.339379 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bccmr" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="registry-server" containerID="cri-o://c20c683f6acdd1ff7a3d66ed39ad63230a4acbb45c2495ec4b7e87ec4b52cb6d" gracePeriod=2 Feb 28 10:41:38 crc kubenswrapper[4972]: I0228 10:41:38.653343 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerID="c20c683f6acdd1ff7a3d66ed39ad63230a4acbb45c2495ec4b7e87ec4b52cb6d" exitCode=0 Feb 28 10:41:38 crc kubenswrapper[4972]: I0228 10:41:38.653416 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerDied","Data":"c20c683f6acdd1ff7a3d66ed39ad63230a4acbb45c2495ec4b7e87ec4b52cb6d"} Feb 28 10:41:38 crc kubenswrapper[4972]: I0228 10:41:38.907279 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.010489 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content\") pod \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.010656 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities\") pod \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.010738 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9dq9\" (UniqueName: \"kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9\") pod \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\" (UID: \"c3023e1b-7bcf-46b1-b61c-92cf0251647e\") " Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.011725 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities" (OuterVolumeSpecName: "utilities") pod "c3023e1b-7bcf-46b1-b61c-92cf0251647e" (UID: "c3023e1b-7bcf-46b1-b61c-92cf0251647e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.016691 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9" (OuterVolumeSpecName: "kube-api-access-m9dq9") pod "c3023e1b-7bcf-46b1-b61c-92cf0251647e" (UID: "c3023e1b-7bcf-46b1-b61c-92cf0251647e"). InnerVolumeSpecName "kube-api-access-m9dq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.037226 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3023e1b-7bcf-46b1-b61c-92cf0251647e" (UID: "c3023e1b-7bcf-46b1-b61c-92cf0251647e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.112843 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.112886 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3023e1b-7bcf-46b1-b61c-92cf0251647e-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.112902 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9dq9\" (UniqueName: \"kubernetes.io/projected/c3023e1b-7bcf-46b1-b61c-92cf0251647e-kube-api-access-m9dq9\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.662343 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bccmr" event={"ID":"c3023e1b-7bcf-46b1-b61c-92cf0251647e","Type":"ContainerDied","Data":"dd9f8b89d6baf1bc65633b3e98f88fd92b66ef355bc37c2fa41ef829199cbb03"} Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.662416 4972 scope.go:117] "RemoveContainer" containerID="c20c683f6acdd1ff7a3d66ed39ad63230a4acbb45c2495ec4b7e87ec4b52cb6d" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.662440 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bccmr" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.690287 4972 scope.go:117] "RemoveContainer" containerID="ca732736e7da1d057ab4a5a806eb7664826d03a02c4ea0418ab3cfe348325ae0" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.706614 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.716963 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bccmr"] Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.719617 4972 scope.go:117] "RemoveContainer" containerID="7e9b621b53bf5eb85267db6b753363f14407c592efb3f1b6505fe05889c82280" Feb 28 10:41:39 crc kubenswrapper[4972]: I0228 10:41:39.797667 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" path="/var/lib/kubelet/pods/c3023e1b-7bcf-46b1-b61c-92cf0251647e/volumes" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.839516 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjxhk"] Feb 28 10:41:42 crc kubenswrapper[4972]: E0228 10:41:42.840028 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="extract-utilities" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.840042 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="extract-utilities" Feb 28 10:41:42 crc kubenswrapper[4972]: E0228 10:41:42.840055 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="registry-server" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.840062 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="registry-server" Feb 28 10:41:42 crc kubenswrapper[4972]: E0228 10:41:42.840072 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="extract-content" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.840078 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="extract-content" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.840171 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3023e1b-7bcf-46b1-b61c-92cf0251647e" containerName="registry-server" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.840560 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.853637 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjxhk"] Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.978558 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90ee640f-57e9-4266-8984-7b44c04b8d90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.978644 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-certificates\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.978705 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-tls\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.978797 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-bound-sa-token\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.978822 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90ee640f-57e9-4266-8984-7b44c04b8d90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.979139 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-trusted-ca\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.979218 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqwrj\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-kube-api-access-lqwrj\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:42 crc kubenswrapper[4972]: I0228 10:41:42.979301 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.005169 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081563 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-trusted-ca\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081661 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqwrj\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-kube-api-access-lqwrj\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90ee640f-57e9-4266-8984-7b44c04b8d90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081796 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-certificates\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081838 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-tls\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081910 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-bound-sa-token\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.081947 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90ee640f-57e9-4266-8984-7b44c04b8d90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.082753 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/90ee640f-57e9-4266-8984-7b44c04b8d90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.083129 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-certificates\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.083449 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90ee640f-57e9-4266-8984-7b44c04b8d90-trusted-ca\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.089192 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/90ee640f-57e9-4266-8984-7b44c04b8d90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.098028 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-registry-tls\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.098599 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-bound-sa-token\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.119711 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqwrj\" (UniqueName: \"kubernetes.io/projected/90ee640f-57e9-4266-8984-7b44c04b8d90-kube-api-access-lqwrj\") pod \"image-registry-66df7c8f76-tjxhk\" (UID: \"90ee640f-57e9-4266-8984-7b44c04b8d90\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.178421 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.653017 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjxhk"] Feb 28 10:41:43 crc kubenswrapper[4972]: I0228 10:41:43.695363 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" event={"ID":"90ee640f-57e9-4266-8984-7b44c04b8d90","Type":"ContainerStarted","Data":"7bd4b43e01e61828c98324fa506a8b72e6d8e836e1f849381ec24287694aba17"} Feb 28 10:41:44 crc kubenswrapper[4972]: I0228 10:41:44.705290 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" event={"ID":"90ee640f-57e9-4266-8984-7b44c04b8d90","Type":"ContainerStarted","Data":"158a4965999f4317e29805db866af70297dc9bc859159ea97a6e3409287b73bb"} Feb 28 10:41:44 crc kubenswrapper[4972]: I0228 10:41:44.705528 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:41:44 crc kubenswrapper[4972]: I0228 10:41:44.733891 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" podStartSLOduration=2.733862613 podStartE2EDuration="2.733862613s" podCreationTimestamp="2026-02-28 10:41:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:41:44.731885803 +0000 UTC m=+361.643893561" watchObservedRunningTime="2026-02-28 10:41:44.733862613 +0000 UTC m=+361.645870371" Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.332688 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.333632 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" podUID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" containerName="controller-manager" containerID="cri-o://8f99513e9c152428d8113ed5e4e8f88c328b4f33fce0c3cf5c9ed97668738280" gracePeriod=30 Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.356725 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.356973 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" podUID="a0ee3613-ba29-4242-8691-2badf68ec79a" containerName="route-controller-manager" containerID="cri-o://74165d85c7fd2b02776e667c504ada86563513f9a73c1aee02cc712d62618b93" gracePeriod=30 Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.747583 4972 generic.go:334] "Generic (PLEG): container finished" podID="a0ee3613-ba29-4242-8691-2badf68ec79a" containerID="74165d85c7fd2b02776e667c504ada86563513f9a73c1aee02cc712d62618b93" exitCode=0 Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.747676 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" event={"ID":"a0ee3613-ba29-4242-8691-2badf68ec79a","Type":"ContainerDied","Data":"74165d85c7fd2b02776e667c504ada86563513f9a73c1aee02cc712d62618b93"} Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.749756 4972 generic.go:334] "Generic (PLEG): container finished" podID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" containerID="8f99513e9c152428d8113ed5e4e8f88c328b4f33fce0c3cf5c9ed97668738280" exitCode=0 Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.749800 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" event={"ID":"cff2eada-b82e-46d9-afc3-8ab91f6ad064","Type":"ContainerDied","Data":"8f99513e9c152428d8113ed5e4e8f88c328b4f33fce0c3cf5c9ed97668738280"} Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.832519 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:41:50 crc kubenswrapper[4972]: I0228 10:41:50.837547 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017132 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert\") pod \"a0ee3613-ba29-4242-8691-2badf68ec79a\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017432 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config\") pod \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017577 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j7ww\" (UniqueName: \"kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww\") pod \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017690 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config\") pod \"a0ee3613-ba29-4242-8691-2badf68ec79a\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017799 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca\") pod \"a0ee3613-ba29-4242-8691-2badf68ec79a\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.017907 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert\") pod \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.018439 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t7vb\" (UniqueName: \"kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb\") pod \"a0ee3613-ba29-4242-8691-2badf68ec79a\" (UID: \"a0ee3613-ba29-4242-8691-2badf68ec79a\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.018576 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles\") pod \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.018697 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca\") pod \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\" (UID: \"cff2eada-b82e-46d9-afc3-8ab91f6ad064\") " Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.018824 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca" (OuterVolumeSpecName: "client-ca") pod "a0ee3613-ba29-4242-8691-2badf68ec79a" (UID: "a0ee3613-ba29-4242-8691-2badf68ec79a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.018895 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config" (OuterVolumeSpecName: "config") pod "cff2eada-b82e-46d9-afc3-8ab91f6ad064" (UID: "cff2eada-b82e-46d9-afc3-8ab91f6ad064"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.019270 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.019391 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.019555 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cff2eada-b82e-46d9-afc3-8ab91f6ad064" (UID: "cff2eada-b82e-46d9-afc3-8ab91f6ad064"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.020024 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca" (OuterVolumeSpecName: "client-ca") pod "cff2eada-b82e-46d9-afc3-8ab91f6ad064" (UID: "cff2eada-b82e-46d9-afc3-8ab91f6ad064"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.020190 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config" (OuterVolumeSpecName: "config") pod "a0ee3613-ba29-4242-8691-2badf68ec79a" (UID: "a0ee3613-ba29-4242-8691-2badf68ec79a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.023398 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cff2eada-b82e-46d9-afc3-8ab91f6ad064" (UID: "cff2eada-b82e-46d9-afc3-8ab91f6ad064"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.024025 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a0ee3613-ba29-4242-8691-2badf68ec79a" (UID: "a0ee3613-ba29-4242-8691-2badf68ec79a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.024259 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww" (OuterVolumeSpecName: "kube-api-access-2j7ww") pod "cff2eada-b82e-46d9-afc3-8ab91f6ad064" (UID: "cff2eada-b82e-46d9-afc3-8ab91f6ad064"). InnerVolumeSpecName "kube-api-access-2j7ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.024856 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb" (OuterVolumeSpecName: "kube-api-access-5t7vb") pod "a0ee3613-ba29-4242-8691-2badf68ec79a" (UID: "a0ee3613-ba29-4242-8691-2badf68ec79a"). InnerVolumeSpecName "kube-api-access-5t7vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120742 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff2eada-b82e-46d9-afc3-8ab91f6ad064-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120825 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t7vb\" (UniqueName: \"kubernetes.io/projected/a0ee3613-ba29-4242-8691-2badf68ec79a-kube-api-access-5t7vb\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120848 4972 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120868 4972 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cff2eada-b82e-46d9-afc3-8ab91f6ad064-client-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120887 4972 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ee3613-ba29-4242-8691-2badf68ec79a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120905 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j7ww\" (UniqueName: \"kubernetes.io/projected/cff2eada-b82e-46d9-afc3-8ab91f6ad064-kube-api-access-2j7ww\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.120924 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ee3613-ba29-4242-8691-2badf68ec79a-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.761790 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.761929 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6" event={"ID":"a0ee3613-ba29-4242-8691-2badf68ec79a","Type":"ContainerDied","Data":"345fefd81d0ff86c4a6ed19556d153d80b91cc0477812c74d8ba149e859de7b2"} Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.762043 4972 scope.go:117] "RemoveContainer" containerID="74165d85c7fd2b02776e667c504ada86563513f9a73c1aee02cc712d62618b93" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.765999 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" event={"ID":"cff2eada-b82e-46d9-afc3-8ab91f6ad064","Type":"ContainerDied","Data":"5f02f5e7acfe7f7a78fbcece5684071b1be17a80642d518a8a932e695afcb9ab"} Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.766074 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-759b4c7f9b-dstl6" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.814325 4972 scope.go:117] "RemoveContainer" containerID="8f99513e9c152428d8113ed5e4e8f88c328b4f33fce0c3cf5c9ed97668738280" Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.839555 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.848174 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5479c9f4cd-lxvd6"] Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.853081 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:41:51 crc kubenswrapper[4972]: I0228 10:41:51.858274 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-759b4c7f9b-dstl6"] Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.195200 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-7nwbt"] Feb 28 10:41:52 crc kubenswrapper[4972]: E0228 10:41:52.195822 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" containerName="controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.195863 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" containerName="controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: E0228 10:41:52.195934 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ee3613-ba29-4242-8691-2badf68ec79a" containerName="route-controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.195952 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ee3613-ba29-4242-8691-2badf68ec79a" containerName="route-controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.196200 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ee3613-ba29-4242-8691-2badf68ec79a" containerName="route-controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.196250 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" containerName="controller-manager" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.197293 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.201973 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.203608 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.203749 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.204653 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.205084 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.205142 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.205340 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4"] Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.209675 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.217979 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.218506 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.219823 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.220143 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.220353 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.220637 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4"] Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.220977 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.225922 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.237747 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-7nwbt"] Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345132 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-config\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345210 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-client-ca\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345237 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jk7\" (UniqueName: \"kubernetes.io/projected/470df015-2ad7-4f50-b4cd-34198dddcf08-kube-api-access-p2jk7\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345269 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470df015-2ad7-4f50-b4cd-34198dddcf08-serving-cert\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345653 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-client-ca\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345754 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b2gj\" (UniqueName: \"kubernetes.io/projected/e169b493-0f30-4c7c-bff3-1454d3756500-kube-api-access-6b2gj\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.345986 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.346041 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-config\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.346085 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e169b493-0f30-4c7c-bff3-1454d3756500-serving-cert\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.447711 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e169b493-0f30-4c7c-bff3-1454d3756500-serving-cert\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.447800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-config\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.447845 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-client-ca\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.447884 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jk7\" (UniqueName: \"kubernetes.io/projected/470df015-2ad7-4f50-b4cd-34198dddcf08-kube-api-access-p2jk7\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.447941 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470df015-2ad7-4f50-b4cd-34198dddcf08-serving-cert\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.448019 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-client-ca\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.448070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b2gj\" (UniqueName: \"kubernetes.io/projected/e169b493-0f30-4c7c-bff3-1454d3756500-kube-api-access-6b2gj\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.448162 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.448198 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-config\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.450262 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-client-ca\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.450427 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-client-ca\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.450590 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/470df015-2ad7-4f50-b4cd-34198dddcf08-config\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.451027 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-config\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.451965 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e169b493-0f30-4c7c-bff3-1454d3756500-proxy-ca-bundles\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.466093 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/470df015-2ad7-4f50-b4cd-34198dddcf08-serving-cert\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.466555 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e169b493-0f30-4c7c-bff3-1454d3756500-serving-cert\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.480137 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b2gj\" (UniqueName: \"kubernetes.io/projected/e169b493-0f30-4c7c-bff3-1454d3756500-kube-api-access-6b2gj\") pod \"controller-manager-f746d79b7-7nwbt\" (UID: \"e169b493-0f30-4c7c-bff3-1454d3756500\") " pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.482407 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jk7\" (UniqueName: \"kubernetes.io/projected/470df015-2ad7-4f50-b4cd-34198dddcf08-kube-api-access-p2jk7\") pod \"route-controller-manager-745bb64d9b-m9kc4\" (UID: \"470df015-2ad7-4f50-b4cd-34198dddcf08\") " pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.529409 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.540993 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.780048 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4"] Feb 28 10:41:52 crc kubenswrapper[4972]: I0228 10:41:52.857281 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f746d79b7-7nwbt"] Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.800043 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ee3613-ba29-4242-8691-2badf68ec79a" path="/var/lib/kubelet/pods/a0ee3613-ba29-4242-8691-2badf68ec79a/volumes" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.801124 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cff2eada-b82e-46d9-afc3-8ab91f6ad064" path="/var/lib/kubelet/pods/cff2eada-b82e-46d9-afc3-8ab91f6ad064/volumes" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.812214 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" event={"ID":"e169b493-0f30-4c7c-bff3-1454d3756500","Type":"ContainerStarted","Data":"43e33dc79376cd25f95f03c328fc06726996e91405bad8e557c08ec34173d6ba"} Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.812268 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" event={"ID":"e169b493-0f30-4c7c-bff3-1454d3756500","Type":"ContainerStarted","Data":"da39c2c256a5b94f90c8b4cbbdcd92a3023336d5b76de2a4acb442a044dceb84"} Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.812488 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.816135 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" event={"ID":"470df015-2ad7-4f50-b4cd-34198dddcf08","Type":"ContainerStarted","Data":"eec662054c54dbd37186ec24cf62790650ed2c70638475f3cf067f1330bf07e3"} Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.816191 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" event={"ID":"470df015-2ad7-4f50-b4cd-34198dddcf08","Type":"ContainerStarted","Data":"56f7375735ea886e1c585deffc315f0f2f580c82ee3f9fa56af3476354fdb5c3"} Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.816382 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.818356 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.821696 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.891117 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-745bb64d9b-m9kc4" podStartSLOduration=3.89110003 podStartE2EDuration="3.89110003s" podCreationTimestamp="2026-02-28 10:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:41:53.890184162 +0000 UTC m=+370.802191900" watchObservedRunningTime="2026-02-28 10:41:53.89110003 +0000 UTC m=+370.803107768" Feb 28 10:41:53 crc kubenswrapper[4972]: I0228 10:41:53.920280 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f746d79b7-7nwbt" podStartSLOduration=3.920259675 podStartE2EDuration="3.920259675s" podCreationTimestamp="2026-02-28 10:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:41:53.918346436 +0000 UTC m=+370.830354174" watchObservedRunningTime="2026-02-28 10:41:53.920259675 +0000 UTC m=+370.832267413" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.165180 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537922-jstkp"] Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.167661 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.175009 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.175776 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.180609 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.181737 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537922-jstkp"] Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.262439 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28tcm\" (UniqueName: \"kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm\") pod \"auto-csr-approver-29537922-jstkp\" (UID: \"12c246d4-f5f0-4d59-8f40-3d891edb2cff\") " pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.364524 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28tcm\" (UniqueName: \"kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm\") pod \"auto-csr-approver-29537922-jstkp\" (UID: \"12c246d4-f5f0-4d59-8f40-3d891edb2cff\") " pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.390568 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28tcm\" (UniqueName: \"kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm\") pod \"auto-csr-approver-29537922-jstkp\" (UID: \"12c246d4-f5f0-4d59-8f40-3d891edb2cff\") " pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.485275 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:00 crc kubenswrapper[4972]: I0228 10:42:00.918506 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537922-jstkp"] Feb 28 10:42:01 crc kubenswrapper[4972]: I0228 10:42:01.870846 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537922-jstkp" event={"ID":"12c246d4-f5f0-4d59-8f40-3d891edb2cff","Type":"ContainerStarted","Data":"24f466033cbbcb2e7325af255830078fb05caf251dfbc864b40104c583167404"} Feb 28 10:42:02 crc kubenswrapper[4972]: I0228 10:42:02.881771 4972 generic.go:334] "Generic (PLEG): container finished" podID="12c246d4-f5f0-4d59-8f40-3d891edb2cff" containerID="2b44db430ad9cb26d45f4ea4a21a594ac3d534096bd21c254cca436dbe975cdd" exitCode=0 Feb 28 10:42:02 crc kubenswrapper[4972]: I0228 10:42:02.881852 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537922-jstkp" event={"ID":"12c246d4-f5f0-4d59-8f40-3d891edb2cff","Type":"ContainerDied","Data":"2b44db430ad9cb26d45f4ea4a21a594ac3d534096bd21c254cca436dbe975cdd"} Feb 28 10:42:03 crc kubenswrapper[4972]: I0228 10:42:03.185024 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tjxhk" Feb 28 10:42:03 crc kubenswrapper[4972]: I0228 10:42:03.248215 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.334856 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.421069 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28tcm\" (UniqueName: \"kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm\") pod \"12c246d4-f5f0-4d59-8f40-3d891edb2cff\" (UID: \"12c246d4-f5f0-4d59-8f40-3d891edb2cff\") " Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.426355 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm" (OuterVolumeSpecName: "kube-api-access-28tcm") pod "12c246d4-f5f0-4d59-8f40-3d891edb2cff" (UID: "12c246d4-f5f0-4d59-8f40-3d891edb2cff"). InnerVolumeSpecName "kube-api-access-28tcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.522973 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28tcm\" (UniqueName: \"kubernetes.io/projected/12c246d4-f5f0-4d59-8f40-3d891edb2cff-kube-api-access-28tcm\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.898741 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537922-jstkp" event={"ID":"12c246d4-f5f0-4d59-8f40-3d891edb2cff","Type":"ContainerDied","Data":"24f466033cbbcb2e7325af255830078fb05caf251dfbc864b40104c583167404"} Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.899094 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24f466033cbbcb2e7325af255830078fb05caf251dfbc864b40104c583167404" Feb 28 10:42:04 crc kubenswrapper[4972]: I0228 10:42:04.898863 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537922-jstkp" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.793597 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.794523 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wzx7z" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="registry-server" containerID="cri-o://42a51a97831116a449d0f1dd1d0d035feafd9a2fb8a851a26796448131519a78" gracePeriod=30 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.810044 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.810301 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6jglk" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="registry-server" containerID="cri-o://fb61785bbe1dfd1d8c33eaa49b10dec9e8523527e2940586abdb909d0bbcd54d" gracePeriod=30 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.815232 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.815449 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerName="marketplace-operator" containerID="cri-o://9c0fed6158775e65627b158bdab548da2b431a7284c8dac657ab4f7484c411be" gracePeriod=30 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.820686 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.820975 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2j69x" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="registry-server" containerID="cri-o://8dc96ea3620ee308b11636e4c5f6a0bde15eed6a3212b4eeffff4759fb731756" gracePeriod=30 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.833549 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m88zk"] Feb 28 10:42:12 crc kubenswrapper[4972]: E0228 10:42:12.833789 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c246d4-f5f0-4d59-8f40-3d891edb2cff" containerName="oc" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.833805 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c246d4-f5f0-4d59-8f40-3d891edb2cff" containerName="oc" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.833915 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c246d4-f5f0-4d59-8f40-3d891edb2cff" containerName="oc" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.834310 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.844539 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.844793 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f45gq" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="registry-server" containerID="cri-o://8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809" gracePeriod=30 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.851974 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m88zk"] Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.874833 4972 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-c6hjj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": read tcp 10.217.0.2:56956->10.217.0.35:8080: read: connection reset by peer" start-of-body= Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.874892 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": read tcp 10.217.0.2:56956->10.217.0.35:8080: read: connection reset by peer" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.944658 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.944733 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.944886 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fqxk\" (UniqueName: \"kubernetes.io/projected/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-kube-api-access-5fqxk\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.965549 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerID="8dc96ea3620ee308b11636e4c5f6a0bde15eed6a3212b4eeffff4759fb731756" exitCode=0 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.965595 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerDied","Data":"8dc96ea3620ee308b11636e4c5f6a0bde15eed6a3212b4eeffff4759fb731756"} Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.971675 4972 generic.go:334] "Generic (PLEG): container finished" podID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerID="fb61785bbe1dfd1d8c33eaa49b10dec9e8523527e2940586abdb909d0bbcd54d" exitCode=0 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.971706 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerDied","Data":"fb61785bbe1dfd1d8c33eaa49b10dec9e8523527e2940586abdb909d0bbcd54d"} Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.982544 4972 generic.go:334] "Generic (PLEG): container finished" podID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerID="9c0fed6158775e65627b158bdab548da2b431a7284c8dac657ab4f7484c411be" exitCode=0 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.982593 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" event={"ID":"0dbb4a63-be87-47b1-8650-59c40ed5f7b5","Type":"ContainerDied","Data":"9c0fed6158775e65627b158bdab548da2b431a7284c8dac657ab4f7484c411be"} Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.984923 4972 generic.go:334] "Generic (PLEG): container finished" podID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerID="42a51a97831116a449d0f1dd1d0d035feafd9a2fb8a851a26796448131519a78" exitCode=0 Feb 28 10:42:12 crc kubenswrapper[4972]: I0228 10:42:12.984956 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerDied","Data":"42a51a97831116a449d0f1dd1d0d035feafd9a2fb8a851a26796448131519a78"} Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.046289 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fqxk\" (UniqueName: \"kubernetes.io/projected/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-kube-api-access-5fqxk\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.046406 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.046480 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.047964 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.052722 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.067165 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fqxk\" (UniqueName: \"kubernetes.io/projected/72e5fba8-6a6e-4310-bcb6-ad064ec74bb9-kube-api-access-5fqxk\") pod \"marketplace-operator-79b997595-m88zk\" (UID: \"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.152696 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.332865 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.452637 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities\") pod \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.453519 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content\") pod \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.453654 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgw7r\" (UniqueName: \"kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r\") pod \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\" (UID: \"285bfdff-92e3-4ec5-bbd8-745860eef1fa\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.458725 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities" (OuterVolumeSpecName: "utilities") pod "285bfdff-92e3-4ec5-bbd8-745860eef1fa" (UID: "285bfdff-92e3-4ec5-bbd8-745860eef1fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.487041 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r" (OuterVolumeSpecName: "kube-api-access-cgw7r") pod "285bfdff-92e3-4ec5-bbd8-745860eef1fa" (UID: "285bfdff-92e3-4ec5-bbd8-745860eef1fa"). InnerVolumeSpecName "kube-api-access-cgw7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.545934 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "285bfdff-92e3-4ec5-bbd8-745860eef1fa" (UID: "285bfdff-92e3-4ec5-bbd8-745860eef1fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.555838 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.555868 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/285bfdff-92e3-4ec5-bbd8-745860eef1fa-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.555880 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgw7r\" (UniqueName: \"kubernetes.io/projected/285bfdff-92e3-4ec5-bbd8-745860eef1fa-kube-api-access-cgw7r\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.665952 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.671722 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.677495 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.685791 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766592 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca\") pod \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766675 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlnmm\" (UniqueName: \"kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm\") pod \"af90cea3-090d-4f18-88e1-c6de1a5fc469\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766712 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgpd7\" (UniqueName: \"kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7\") pod \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766747 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics\") pod \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766779 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content\") pod \"af90cea3-090d-4f18-88e1-c6de1a5fc469\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766807 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzjp6\" (UniqueName: \"kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6\") pod \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766827 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmhbq\" (UniqueName: \"kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq\") pod \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\" (UID: \"0dbb4a63-be87-47b1-8650-59c40ed5f7b5\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766861 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities\") pod \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766878 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content\") pod \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\" (UID: \"ad7908e4-f813-4c17-a672-a9ff32bdfd8c\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766900 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities\") pod \"af90cea3-090d-4f18-88e1-c6de1a5fc469\" (UID: \"af90cea3-090d-4f18-88e1-c6de1a5fc469\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766914 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content\") pod \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.766937 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities\") pod \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\" (UID: \"66ead19f-03ec-41d1-82b4-9697bbb1e6a6\") " Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.767755 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities" (OuterVolumeSpecName: "utilities") pod "66ead19f-03ec-41d1-82b4-9697bbb1e6a6" (UID: "66ead19f-03ec-41d1-82b4-9697bbb1e6a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.768612 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0dbb4a63-be87-47b1-8650-59c40ed5f7b5" (UID: "0dbb4a63-be87-47b1-8650-59c40ed5f7b5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.769310 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities" (OuterVolumeSpecName: "utilities") pod "ad7908e4-f813-4c17-a672-a9ff32bdfd8c" (UID: "ad7908e4-f813-4c17-a672-a9ff32bdfd8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.769413 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities" (OuterVolumeSpecName: "utilities") pod "af90cea3-090d-4f18-88e1-c6de1a5fc469" (UID: "af90cea3-090d-4f18-88e1-c6de1a5fc469"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.771854 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6" (OuterVolumeSpecName: "kube-api-access-hzjp6") pod "ad7908e4-f813-4c17-a672-a9ff32bdfd8c" (UID: "ad7908e4-f813-4c17-a672-a9ff32bdfd8c"). InnerVolumeSpecName "kube-api-access-hzjp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.772407 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0dbb4a63-be87-47b1-8650-59c40ed5f7b5" (UID: "0dbb4a63-be87-47b1-8650-59c40ed5f7b5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.773287 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7" (OuterVolumeSpecName: "kube-api-access-mgpd7") pod "66ead19f-03ec-41d1-82b4-9697bbb1e6a6" (UID: "66ead19f-03ec-41d1-82b4-9697bbb1e6a6"). InnerVolumeSpecName "kube-api-access-mgpd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.778002 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm" (OuterVolumeSpecName: "kube-api-access-rlnmm") pod "af90cea3-090d-4f18-88e1-c6de1a5fc469" (UID: "af90cea3-090d-4f18-88e1-c6de1a5fc469"). InnerVolumeSpecName "kube-api-access-rlnmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.778902 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq" (OuterVolumeSpecName: "kube-api-access-rmhbq") pod "0dbb4a63-be87-47b1-8650-59c40ed5f7b5" (UID: "0dbb4a63-be87-47b1-8650-59c40ed5f7b5"). InnerVolumeSpecName "kube-api-access-rmhbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.802275 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad7908e4-f813-4c17-a672-a9ff32bdfd8c" (UID: "ad7908e4-f813-4c17-a672-a9ff32bdfd8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.866643 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m88zk"] Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.866843 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66ead19f-03ec-41d1-82b4-9697bbb1e6a6" (UID: "66ead19f-03ec-41d1-82b4-9697bbb1e6a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868274 4972 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868298 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzjp6\" (UniqueName: \"kubernetes.io/projected/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-kube-api-access-hzjp6\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868307 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmhbq\" (UniqueName: \"kubernetes.io/projected/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-kube-api-access-rmhbq\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868318 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868329 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7908e4-f813-4c17-a672-a9ff32bdfd8c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868337 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868345 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868353 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868360 4972 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0dbb4a63-be87-47b1-8650-59c40ed5f7b5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868369 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlnmm\" (UniqueName: \"kubernetes.io/projected/af90cea3-090d-4f18-88e1-c6de1a5fc469-kube-api-access-rlnmm\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.868377 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgpd7\" (UniqueName: \"kubernetes.io/projected/66ead19f-03ec-41d1-82b4-9697bbb1e6a6-kube-api-access-mgpd7\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.926288 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af90cea3-090d-4f18-88e1-c6de1a5fc469" (UID: "af90cea3-090d-4f18-88e1-c6de1a5fc469"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.970045 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af90cea3-090d-4f18-88e1-c6de1a5fc469-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.993703 4972 generic.go:334] "Generic (PLEG): container finished" podID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerID="8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809" exitCode=0 Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.993763 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerDied","Data":"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809"} Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.993824 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f45gq" event={"ID":"af90cea3-090d-4f18-88e1-c6de1a5fc469","Type":"ContainerDied","Data":"799df91914a3dbe4e458d7c17041ff5ba28f9ca03f5299ff858b3cd47c2a5380"} Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.993845 4972 scope.go:117] "RemoveContainer" containerID="8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.994001 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f45gq" Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.997670 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" event={"ID":"0dbb4a63-be87-47b1-8650-59c40ed5f7b5","Type":"ContainerDied","Data":"7b41de8b1b785cc99ce8e49a93ee815446c0e770fc1d1933c063f1eaf9265cd9"} Feb 28 10:42:13 crc kubenswrapper[4972]: I0228 10:42:13.998039 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6hjj" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.003757 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wzx7z" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.003787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wzx7z" event={"ID":"285bfdff-92e3-4ec5-bbd8-745860eef1fa","Type":"ContainerDied","Data":"e8159e1c553323ded7f90e7db8a6d395cadfddbd50c736eed4227a737f84af38"} Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.009663 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j69x" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.009657 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j69x" event={"ID":"ad7908e4-f813-4c17-a672-a9ff32bdfd8c","Type":"ContainerDied","Data":"8ddce064c8404db9141d14e0c2464f891d20eaf08b073a0552568f8d9b323c71"} Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.013747 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" event={"ID":"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9","Type":"ContainerStarted","Data":"c144573cf0ede7259edde6889095b032a99acf30e353f381a18b929c769cb2e3"} Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.017695 4972 scope.go:117] "RemoveContainer" containerID="3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.022136 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jglk" event={"ID":"66ead19f-03ec-41d1-82b4-9697bbb1e6a6","Type":"ContainerDied","Data":"5227ff53a47d49168ed6a2712f2e0360ae1028ceabcfadf2bdfee23db3dcab65"} Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.022257 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jglk" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.026273 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.036834 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6hjj"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.045773 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.049929 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f45gq"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.057301 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.069707 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wzx7z"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.069683 4972 scope.go:117] "RemoveContainer" containerID="6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.070904 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.075853 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j69x"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.090063 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.094069 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6jglk"] Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.098272 4972 scope.go:117] "RemoveContainer" containerID="8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809" Feb 28 10:42:14 crc kubenswrapper[4972]: E0228 10:42:14.098675 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809\": container with ID starting with 8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809 not found: ID does not exist" containerID="8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.098784 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809"} err="failed to get container status \"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809\": rpc error: code = NotFound desc = could not find container \"8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809\": container with ID starting with 8b21cb38491d5a569ac61feaa5bd309f1b679bbc21b3fd72e9b61eafeb36b809 not found: ID does not exist" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.098874 4972 scope.go:117] "RemoveContainer" containerID="3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257" Feb 28 10:42:14 crc kubenswrapper[4972]: E0228 10:42:14.099188 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257\": container with ID starting with 3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257 not found: ID does not exist" containerID="3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.099278 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257"} err="failed to get container status \"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257\": rpc error: code = NotFound desc = could not find container \"3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257\": container with ID starting with 3711ec3de104d7c0515edaae6e26a6143146c3cee879d9855cb6f0ff377d7257 not found: ID does not exist" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.099348 4972 scope.go:117] "RemoveContainer" containerID="6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665" Feb 28 10:42:14 crc kubenswrapper[4972]: E0228 10:42:14.099674 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665\": container with ID starting with 6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665 not found: ID does not exist" containerID="6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.099764 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665"} err="failed to get container status \"6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665\": rpc error: code = NotFound desc = could not find container \"6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665\": container with ID starting with 6b1b586fb4a77cd2fdeb9238ae19d14dba5e0466fdf253527cbab5bf8ae10665 not found: ID does not exist" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.099830 4972 scope.go:117] "RemoveContainer" containerID="9c0fed6158775e65627b158bdab548da2b431a7284c8dac657ab4f7484c411be" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.125637 4972 scope.go:117] "RemoveContainer" containerID="42a51a97831116a449d0f1dd1d0d035feafd9a2fb8a851a26796448131519a78" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.145907 4972 scope.go:117] "RemoveContainer" containerID="8187d38cae4b15655eb3003e03926b8cc107a474bc480a2198928a802dd7dc0c" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.164167 4972 scope.go:117] "RemoveContainer" containerID="b038f7d2f38eb1192a0087044f7253e8d0d7e66580b84ee41835d76c6c071ecd" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.184046 4972 scope.go:117] "RemoveContainer" containerID="8dc96ea3620ee308b11636e4c5f6a0bde15eed6a3212b4eeffff4759fb731756" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.204679 4972 scope.go:117] "RemoveContainer" containerID="7927e77085602432e3c47acb85b6653a07d802ba23fa800f7c891f923abd27ce" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.225280 4972 scope.go:117] "RemoveContainer" containerID="d084c942847d9f0e5ebd73330e65697132aa4cc6c742ba91ecf668de5c4a3397" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.241534 4972 scope.go:117] "RemoveContainer" containerID="fb61785bbe1dfd1d8c33eaa49b10dec9e8523527e2940586abdb909d0bbcd54d" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.257374 4972 scope.go:117] "RemoveContainer" containerID="f962e25bab637209452d9ce7a4762fae072d2350ff586d7d2fcf502374f44562" Feb 28 10:42:14 crc kubenswrapper[4972]: I0228 10:42:14.274841 4972 scope.go:117] "RemoveContainer" containerID="439742b31ac32f4e1d2846a1d3bcaeb86f15fb8a51aaf6af18a7ba949f5cd3c2" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.008765 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2ggdk"] Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009307 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009321 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009332 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009338 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009349 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009355 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009364 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009370 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009376 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerName="marketplace-operator" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009382 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerName="marketplace-operator" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009393 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009399 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009406 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009415 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009423 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009428 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009437 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009442 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="extract-content" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009450 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009460 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009489 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009495 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009502 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009508 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: E0228 10:42:15.009516 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009521 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="extract-utilities" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009622 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009633 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" containerName="marketplace-operator" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009645 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009656 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.009685 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" containerName="registry-server" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.010525 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.015113 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.019946 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ggdk"] Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.038541 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" event={"ID":"72e5fba8-6a6e-4310-bcb6-ad064ec74bb9","Type":"ContainerStarted","Data":"087decd25638ab932a9e8e058c2a7155dab6c961bc4cd3646a077374b967759e"} Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.039605 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.042647 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.058918 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m88zk" podStartSLOduration=3.058898312 podStartE2EDuration="3.058898312s" podCreationTimestamp="2026-02-28 10:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:42:15.055802858 +0000 UTC m=+391.967810606" watchObservedRunningTime="2026-02-28 10:42:15.058898312 +0000 UTC m=+391.970906050" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.084567 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qddln\" (UniqueName: \"kubernetes.io/projected/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-kube-api-access-qddln\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.084671 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-utilities\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.084711 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-catalog-content\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.186414 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qddln\" (UniqueName: \"kubernetes.io/projected/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-kube-api-access-qddln\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.186606 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-utilities\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.186633 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-catalog-content\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.187373 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-catalog-content\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.187710 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-utilities\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.212423 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jzl5b"] Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.214813 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.218894 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.220254 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jzl5b"] Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.223304 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qddln\" (UniqueName: \"kubernetes.io/projected/0c340c42-4bb5-4395-a1cf-2cbdcd29f894-kube-api-access-qddln\") pod \"redhat-marketplace-2ggdk\" (UID: \"0c340c42-4bb5-4395-a1cf-2cbdcd29f894\") " pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.288423 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-catalog-content\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.288497 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-utilities\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.288586 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfgg\" (UniqueName: \"kubernetes.io/projected/c549bc93-54c0-4ead-95d0-3b94f86af486-kube-api-access-ppfgg\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.338636 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.390449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-catalog-content\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.390567 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-utilities\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.390713 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfgg\" (UniqueName: \"kubernetes.io/projected/c549bc93-54c0-4ead-95d0-3b94f86af486-kube-api-access-ppfgg\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.391454 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-utilities\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.391534 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c549bc93-54c0-4ead-95d0-3b94f86af486-catalog-content\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.410473 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfgg\" (UniqueName: \"kubernetes.io/projected/c549bc93-54c0-4ead-95d0-3b94f86af486-kube-api-access-ppfgg\") pod \"community-operators-jzl5b\" (UID: \"c549bc93-54c0-4ead-95d0-3b94f86af486\") " pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.547043 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:15 crc kubenswrapper[4972]: W0228 10:42:15.804989 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c340c42_4bb5_4395_a1cf_2cbdcd29f894.slice/crio-9eb0ad9c8088bfd663339814c9f1199d0dc78022f3bf79759e32144f9a428c4f WatchSource:0}: Error finding container 9eb0ad9c8088bfd663339814c9f1199d0dc78022f3bf79759e32144f9a428c4f: Status 404 returned error can't find the container with id 9eb0ad9c8088bfd663339814c9f1199d0dc78022f3bf79759e32144f9a428c4f Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.805537 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dbb4a63-be87-47b1-8650-59c40ed5f7b5" path="/var/lib/kubelet/pods/0dbb4a63-be87-47b1-8650-59c40ed5f7b5/volumes" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.806796 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285bfdff-92e3-4ec5-bbd8-745860eef1fa" path="/var/lib/kubelet/pods/285bfdff-92e3-4ec5-bbd8-745860eef1fa/volumes" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.807844 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66ead19f-03ec-41d1-82b4-9697bbb1e6a6" path="/var/lib/kubelet/pods/66ead19f-03ec-41d1-82b4-9697bbb1e6a6/volumes" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.809496 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7908e4-f813-4c17-a672-a9ff32bdfd8c" path="/var/lib/kubelet/pods/ad7908e4-f813-4c17-a672-a9ff32bdfd8c/volumes" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.810968 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af90cea3-090d-4f18-88e1-c6de1a5fc469" path="/var/lib/kubelet/pods/af90cea3-090d-4f18-88e1-c6de1a5fc469/volumes" Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.811967 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ggdk"] Feb 28 10:42:15 crc kubenswrapper[4972]: I0228 10:42:15.943514 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jzl5b"] Feb 28 10:42:15 crc kubenswrapper[4972]: W0228 10:42:15.951502 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc549bc93_54c0_4ead_95d0_3b94f86af486.slice/crio-1d398c0d12e5c6216c2cc741b605d21e67ca9462eeb14b80e8cb3b003bb529c6 WatchSource:0}: Error finding container 1d398c0d12e5c6216c2cc741b605d21e67ca9462eeb14b80e8cb3b003bb529c6: Status 404 returned error can't find the container with id 1d398c0d12e5c6216c2cc741b605d21e67ca9462eeb14b80e8cb3b003bb529c6 Feb 28 10:42:16 crc kubenswrapper[4972]: I0228 10:42:16.051764 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jzl5b" event={"ID":"c549bc93-54c0-4ead-95d0-3b94f86af486","Type":"ContainerStarted","Data":"1d398c0d12e5c6216c2cc741b605d21e67ca9462eeb14b80e8cb3b003bb529c6"} Feb 28 10:42:16 crc kubenswrapper[4972]: I0228 10:42:16.053664 4972 generic.go:334] "Generic (PLEG): container finished" podID="0c340c42-4bb5-4395-a1cf-2cbdcd29f894" containerID="9710f51df0e14e70bb18f7fee92e23d6f7b826ec6fd6fada02ab2018e095fb80" exitCode=0 Feb 28 10:42:16 crc kubenswrapper[4972]: I0228 10:42:16.053786 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ggdk" event={"ID":"0c340c42-4bb5-4395-a1cf-2cbdcd29f894","Type":"ContainerDied","Data":"9710f51df0e14e70bb18f7fee92e23d6f7b826ec6fd6fada02ab2018e095fb80"} Feb 28 10:42:16 crc kubenswrapper[4972]: I0228 10:42:16.053844 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ggdk" event={"ID":"0c340c42-4bb5-4395-a1cf-2cbdcd29f894","Type":"ContainerStarted","Data":"9eb0ad9c8088bfd663339814c9f1199d0dc78022f3bf79759e32144f9a428c4f"} Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.062865 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ggdk" event={"ID":"0c340c42-4bb5-4395-a1cf-2cbdcd29f894","Type":"ContainerStarted","Data":"484006af80338a42c4e1f03ec7644f19754b1be01305abb8d4fde643d511dfc5"} Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.064590 4972 generic.go:334] "Generic (PLEG): container finished" podID="c549bc93-54c0-4ead-95d0-3b94f86af486" containerID="61f351e512c685a3f50d412a9ac4396a355923324a64d1a65a23f289275d0db7" exitCode=0 Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.064774 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jzl5b" event={"ID":"c549bc93-54c0-4ead-95d0-3b94f86af486","Type":"ContainerDied","Data":"61f351e512c685a3f50d412a9ac4396a355923324a64d1a65a23f289275d0db7"} Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.408432 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-825kc"] Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.409752 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.415299 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.426073 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-825kc"] Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.519944 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-catalog-content\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.520054 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-utilities\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.520136 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl889\" (UniqueName: \"kubernetes.io/projected/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-kube-api-access-dl889\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.608581 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2sr94"] Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.610325 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.613284 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.624913 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl889\" (UniqueName: \"kubernetes.io/projected/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-kube-api-access-dl889\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.625003 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-catalog-content\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.625067 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-utilities\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.625589 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-utilities\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.626172 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-catalog-content\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.635447 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sr94"] Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.664026 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl889\" (UniqueName: \"kubernetes.io/projected/0d7d47c5-06cc-48b9-b7f4-836f7bffc194-kube-api-access-dl889\") pod \"redhat-operators-825kc\" (UID: \"0d7d47c5-06cc-48b9-b7f4-836f7bffc194\") " pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.726596 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-catalog-content\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.726655 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9hbz\" (UniqueName: \"kubernetes.io/projected/1bd5b0fb-5083-4eff-8225-02ba27a56628-kube-api-access-k9hbz\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.726757 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-utilities\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.730983 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.828351 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-utilities\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.828800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-catalog-content\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.828828 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9hbz\" (UniqueName: \"kubernetes.io/projected/1bd5b0fb-5083-4eff-8225-02ba27a56628-kube-api-access-k9hbz\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.829924 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-utilities\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.830422 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bd5b0fb-5083-4eff-8225-02ba27a56628-catalog-content\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.852637 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9hbz\" (UniqueName: \"kubernetes.io/projected/1bd5b0fb-5083-4eff-8225-02ba27a56628-kube-api-access-k9hbz\") pod \"certified-operators-2sr94\" (UID: \"1bd5b0fb-5083-4eff-8225-02ba27a56628\") " pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:17 crc kubenswrapper[4972]: I0228 10:42:17.950109 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:18 crc kubenswrapper[4972]: I0228 10:42:18.078159 4972 generic.go:334] "Generic (PLEG): container finished" podID="0c340c42-4bb5-4395-a1cf-2cbdcd29f894" containerID="484006af80338a42c4e1f03ec7644f19754b1be01305abb8d4fde643d511dfc5" exitCode=0 Feb 28 10:42:18 crc kubenswrapper[4972]: I0228 10:42:18.078690 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ggdk" event={"ID":"0c340c42-4bb5-4395-a1cf-2cbdcd29f894","Type":"ContainerDied","Data":"484006af80338a42c4e1f03ec7644f19754b1be01305abb8d4fde643d511dfc5"} Feb 28 10:42:18 crc kubenswrapper[4972]: I0228 10:42:18.085606 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jzl5b" event={"ID":"c549bc93-54c0-4ead-95d0-3b94f86af486","Type":"ContainerStarted","Data":"68759b70a276d7e0102545bb76ebd9706721a9890980a6882db2ac3807e0a607"} Feb 28 10:42:18 crc kubenswrapper[4972]: I0228 10:42:18.167918 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-825kc"] Feb 28 10:42:18 crc kubenswrapper[4972]: I0228 10:42:18.340443 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sr94"] Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.094336 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ggdk" event={"ID":"0c340c42-4bb5-4395-a1cf-2cbdcd29f894","Type":"ContainerStarted","Data":"b881c29d57bf9b3c4b4d56b255534c72a5e5cb0f49f07344b0932506418ce171"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.098102 4972 generic.go:334] "Generic (PLEG): container finished" podID="c549bc93-54c0-4ead-95d0-3b94f86af486" containerID="68759b70a276d7e0102545bb76ebd9706721a9890980a6882db2ac3807e0a607" exitCode=0 Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.098170 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jzl5b" event={"ID":"c549bc93-54c0-4ead-95d0-3b94f86af486","Type":"ContainerDied","Data":"68759b70a276d7e0102545bb76ebd9706721a9890980a6882db2ac3807e0a607"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.102216 4972 generic.go:334] "Generic (PLEG): container finished" podID="1bd5b0fb-5083-4eff-8225-02ba27a56628" containerID="82896ee3a40f60e1d38f7f93b9f136a154a7e7830a46f76329ce434f1ea2a17b" exitCode=0 Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.102288 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sr94" event={"ID":"1bd5b0fb-5083-4eff-8225-02ba27a56628","Type":"ContainerDied","Data":"82896ee3a40f60e1d38f7f93b9f136a154a7e7830a46f76329ce434f1ea2a17b"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.102311 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sr94" event={"ID":"1bd5b0fb-5083-4eff-8225-02ba27a56628","Type":"ContainerStarted","Data":"90c84411eddff82600511bc011b07ad23e5d5bfb5d7f098ff5d69d9194da8a5f"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.105136 4972 generic.go:334] "Generic (PLEG): container finished" podID="0d7d47c5-06cc-48b9-b7f4-836f7bffc194" containerID="2274f6d8dabc3a8c836c8c39e6f855d69f922f7663e18ea41acf397a43b59b8e" exitCode=0 Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.105176 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825kc" event={"ID":"0d7d47c5-06cc-48b9-b7f4-836f7bffc194","Type":"ContainerDied","Data":"2274f6d8dabc3a8c836c8c39e6f855d69f922f7663e18ea41acf397a43b59b8e"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.105197 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825kc" event={"ID":"0d7d47c5-06cc-48b9-b7f4-836f7bffc194","Type":"ContainerStarted","Data":"d1e3c04f82af95430c965626ec54a50756a620c49208eb4e55c0570632fb262b"} Feb 28 10:42:19 crc kubenswrapper[4972]: I0228 10:42:19.118921 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2ggdk" podStartSLOduration=2.3517937890000002 podStartE2EDuration="5.118905616s" podCreationTimestamp="2026-02-28 10:42:14 +0000 UTC" firstStartedPulling="2026-02-28 10:42:16.055506639 +0000 UTC m=+392.967514377" lastFinishedPulling="2026-02-28 10:42:18.822618456 +0000 UTC m=+395.734626204" observedRunningTime="2026-02-28 10:42:19.116546334 +0000 UTC m=+396.028554082" watchObservedRunningTime="2026-02-28 10:42:19.118905616 +0000 UTC m=+396.030913354" Feb 28 10:42:20 crc kubenswrapper[4972]: I0228 10:42:20.114348 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sr94" event={"ID":"1bd5b0fb-5083-4eff-8225-02ba27a56628","Type":"ContainerStarted","Data":"a127401abad8e7ab3163e0056cfad36e55571c980e8971ce22140314e2229919"} Feb 28 10:42:20 crc kubenswrapper[4972]: I0228 10:42:20.118163 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825kc" event={"ID":"0d7d47c5-06cc-48b9-b7f4-836f7bffc194","Type":"ContainerStarted","Data":"661a58e5cc547f5e3158d29f7f2e2e2d477473f0cfbc6d63a22c5b194afa4946"} Feb 28 10:42:20 crc kubenswrapper[4972]: I0228 10:42:20.124452 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jzl5b" event={"ID":"c549bc93-54c0-4ead-95d0-3b94f86af486","Type":"ContainerStarted","Data":"a542d688e7262bdaef6708d0d2457a317877573a38fae8551aed5ce21cff4b84"} Feb 28 10:42:20 crc kubenswrapper[4972]: I0228 10:42:20.169934 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jzl5b" podStartSLOduration=2.754478185 podStartE2EDuration="5.169906246s" podCreationTimestamp="2026-02-28 10:42:15 +0000 UTC" firstStartedPulling="2026-02-28 10:42:17.065831143 +0000 UTC m=+393.977838881" lastFinishedPulling="2026-02-28 10:42:19.481259194 +0000 UTC m=+396.393266942" observedRunningTime="2026-02-28 10:42:20.167616436 +0000 UTC m=+397.079624174" watchObservedRunningTime="2026-02-28 10:42:20.169906246 +0000 UTC m=+397.081913974" Feb 28 10:42:21 crc kubenswrapper[4972]: I0228 10:42:21.133580 4972 generic.go:334] "Generic (PLEG): container finished" podID="0d7d47c5-06cc-48b9-b7f4-836f7bffc194" containerID="661a58e5cc547f5e3158d29f7f2e2e2d477473f0cfbc6d63a22c5b194afa4946" exitCode=0 Feb 28 10:42:21 crc kubenswrapper[4972]: I0228 10:42:21.133647 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825kc" event={"ID":"0d7d47c5-06cc-48b9-b7f4-836f7bffc194","Type":"ContainerDied","Data":"661a58e5cc547f5e3158d29f7f2e2e2d477473f0cfbc6d63a22c5b194afa4946"} Feb 28 10:42:21 crc kubenswrapper[4972]: I0228 10:42:21.138079 4972 generic.go:334] "Generic (PLEG): container finished" podID="1bd5b0fb-5083-4eff-8225-02ba27a56628" containerID="a127401abad8e7ab3163e0056cfad36e55571c980e8971ce22140314e2229919" exitCode=0 Feb 28 10:42:21 crc kubenswrapper[4972]: I0228 10:42:21.138138 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sr94" event={"ID":"1bd5b0fb-5083-4eff-8225-02ba27a56628","Type":"ContainerDied","Data":"a127401abad8e7ab3163e0056cfad36e55571c980e8971ce22140314e2229919"} Feb 28 10:42:22 crc kubenswrapper[4972]: I0228 10:42:22.147620 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-825kc" event={"ID":"0d7d47c5-06cc-48b9-b7f4-836f7bffc194","Type":"ContainerStarted","Data":"9ce6529adc9932c4491e3bf75604f8939ca63404ffc8b75da72949cf6929ee81"} Feb 28 10:42:22 crc kubenswrapper[4972]: I0228 10:42:22.152450 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sr94" event={"ID":"1bd5b0fb-5083-4eff-8225-02ba27a56628","Type":"ContainerStarted","Data":"6d63af5bb3a4acda5184a9e1b1b17e8fd4bc56285ec85fe435228c49ee99ff49"} Feb 28 10:42:22 crc kubenswrapper[4972]: I0228 10:42:22.230084 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-825kc" podStartSLOduration=2.818141058 podStartE2EDuration="5.230058724s" podCreationTimestamp="2026-02-28 10:42:17 +0000 UTC" firstStartedPulling="2026-02-28 10:42:19.106977732 +0000 UTC m=+396.018985470" lastFinishedPulling="2026-02-28 10:42:21.518895398 +0000 UTC m=+398.430903136" observedRunningTime="2026-02-28 10:42:22.203793676 +0000 UTC m=+399.115801424" watchObservedRunningTime="2026-02-28 10:42:22.230058724 +0000 UTC m=+399.142066462" Feb 28 10:42:22 crc kubenswrapper[4972]: I0228 10:42:22.232845 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2sr94" podStartSLOduration=2.726072419 podStartE2EDuration="5.232835678s" podCreationTimestamp="2026-02-28 10:42:17 +0000 UTC" firstStartedPulling="2026-02-28 10:42:19.107582201 +0000 UTC m=+396.019589939" lastFinishedPulling="2026-02-28 10:42:21.61434545 +0000 UTC m=+398.526353198" observedRunningTime="2026-02-28 10:42:22.22824772 +0000 UTC m=+399.140255478" watchObservedRunningTime="2026-02-28 10:42:22.232835678 +0000 UTC m=+399.144843416" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.338968 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.339480 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.410907 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.547353 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.547428 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:25 crc kubenswrapper[4972]: I0228 10:42:25.615289 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:26 crc kubenswrapper[4972]: I0228 10:42:26.232968 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2ggdk" Feb 28 10:42:26 crc kubenswrapper[4972]: I0228 10:42:26.259129 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jzl5b" Feb 28 10:42:27 crc kubenswrapper[4972]: I0228 10:42:27.731689 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:27 crc kubenswrapper[4972]: I0228 10:42:27.731750 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:27 crc kubenswrapper[4972]: I0228 10:42:27.951208 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:27 crc kubenswrapper[4972]: I0228 10:42:27.951654 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:28 crc kubenswrapper[4972]: I0228 10:42:28.001329 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:28 crc kubenswrapper[4972]: I0228 10:42:28.260783 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2sr94" Feb 28 10:42:28 crc kubenswrapper[4972]: I0228 10:42:28.312604 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" podUID="8bc8f405-1a09-41db-9582-091d232da716" containerName="registry" containerID="cri-o://106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46" gracePeriod=30 Feb 28 10:42:28 crc kubenswrapper[4972]: I0228 10:42:28.786849 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-825kc" podUID="0d7d47c5-06cc-48b9-b7f4-836f7bffc194" containerName="registry-server" probeResult="failure" output=< Feb 28 10:42:28 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:42:28 crc kubenswrapper[4972]: > Feb 28 10:42:28 crc kubenswrapper[4972]: I0228 10:42:28.982123 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107559 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107677 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107733 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107790 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107824 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107888 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107929 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.107975 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz2nb\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb\") pod \"8bc8f405-1a09-41db-9582-091d232da716\" (UID: \"8bc8f405-1a09-41db-9582-091d232da716\") " Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.110068 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.110244 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.115895 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.116992 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.117548 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb" (OuterVolumeSpecName: "kube-api-access-pz2nb") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "kube-api-access-pz2nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.118117 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.118898 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.144824 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8bc8f405-1a09-41db-9582-091d232da716" (UID: "8bc8f405-1a09-41db-9582-091d232da716"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.202801 4972 generic.go:334] "Generic (PLEG): container finished" podID="8bc8f405-1a09-41db-9582-091d232da716" containerID="106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46" exitCode=0 Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.202878 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.202974 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" event={"ID":"8bc8f405-1a09-41db-9582-091d232da716","Type":"ContainerDied","Data":"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46"} Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.203069 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f7g8b" event={"ID":"8bc8f405-1a09-41db-9582-091d232da716","Type":"ContainerDied","Data":"19adb258528068a3b37686fc0be619fbf16b396c7a729303015a0d00f064b4b9"} Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.203098 4972 scope.go:117] "RemoveContainer" containerID="106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209162 4972 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209204 4972 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209224 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz2nb\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-kube-api-access-pz2nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209238 4972 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f405-1a09-41db-9582-091d232da716-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209251 4972 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc8f405-1a09-41db-9582-091d232da716-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209265 4972 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8bc8f405-1a09-41db-9582-091d232da716-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.209277 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bc8f405-1a09-41db-9582-091d232da716-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.229890 4972 scope.go:117] "RemoveContainer" containerID="106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46" Feb 28 10:42:29 crc kubenswrapper[4972]: E0228 10:42:29.230562 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46\": container with ID starting with 106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46 not found: ID does not exist" containerID="106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.230628 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46"} err="failed to get container status \"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46\": rpc error: code = NotFound desc = could not find container \"106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46\": container with ID starting with 106b1d7c62b40b40f918226fe5618f4669224a95cf414e328c353cf0a7609f46 not found: ID does not exist" Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.248806 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.257296 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f7g8b"] Feb 28 10:42:29 crc kubenswrapper[4972]: I0228 10:42:29.797075 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bc8f405-1a09-41db-9582-091d232da716" path="/var/lib/kubelet/pods/8bc8f405-1a09-41db-9582-091d232da716/volumes" Feb 28 10:42:37 crc kubenswrapper[4972]: I0228 10:42:37.806904 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:42:37 crc kubenswrapper[4972]: I0228 10:42:37.882001 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-825kc" Feb 28 10:43:28 crc kubenswrapper[4972]: I0228 10:43:28.890534 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:43:28 crc kubenswrapper[4972]: I0228 10:43:28.891279 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:43:56 crc kubenswrapper[4972]: I0228 10:43:56.846586 4972 scope.go:117] "RemoveContainer" containerID="571c5d16bcd3380361cd3f5eaf9df084a384ecad6c6df862c61b54e19b0b41a2" Feb 28 10:43:58 crc kubenswrapper[4972]: I0228 10:43:58.890503 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:43:58 crc kubenswrapper[4972]: I0228 10:43:58.891380 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.134082 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537924-h2gm7"] Feb 28 10:44:00 crc kubenswrapper[4972]: E0228 10:44:00.136514 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc8f405-1a09-41db-9582-091d232da716" containerName="registry" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.139938 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc8f405-1a09-41db-9582-091d232da716" containerName="registry" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.140230 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc8f405-1a09-41db-9582-091d232da716" containerName="registry" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.140828 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.143036 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.143378 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.143431 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.149538 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537924-h2gm7"] Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.183423 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4cmx\" (UniqueName: \"kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx\") pod \"auto-csr-approver-29537924-h2gm7\" (UID: \"1fcdc4df-a299-4e3c-970b-47c4e01aa51e\") " pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.284225 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4cmx\" (UniqueName: \"kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx\") pod \"auto-csr-approver-29537924-h2gm7\" (UID: \"1fcdc4df-a299-4e3c-970b-47c4e01aa51e\") " pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.312083 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4cmx\" (UniqueName: \"kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx\") pod \"auto-csr-approver-29537924-h2gm7\" (UID: \"1fcdc4df-a299-4e3c-970b-47c4e01aa51e\") " pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.500207 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.736214 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537924-h2gm7"] Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.755044 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 10:44:00 crc kubenswrapper[4972]: I0228 10:44:00.826643 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" event={"ID":"1fcdc4df-a299-4e3c-970b-47c4e01aa51e","Type":"ContainerStarted","Data":"c53cd232b006f58c9e1379f143519eb4cf288e867374d44ae4bbab471645de75"} Feb 28 10:44:01 crc kubenswrapper[4972]: I0228 10:44:01.834383 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" event={"ID":"1fcdc4df-a299-4e3c-970b-47c4e01aa51e","Type":"ContainerStarted","Data":"9fe3be65c4e68afcb7b9dfa412fad3a1caf9de652a625e28d6a44609f7d5f336"} Feb 28 10:44:01 crc kubenswrapper[4972]: I0228 10:44:01.850599 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" podStartSLOduration=1.065084249 podStartE2EDuration="1.850579443s" podCreationTimestamp="2026-02-28 10:44:00 +0000 UTC" firstStartedPulling="2026-02-28 10:44:00.754781455 +0000 UTC m=+497.666789203" lastFinishedPulling="2026-02-28 10:44:01.540276659 +0000 UTC m=+498.452284397" observedRunningTime="2026-02-28 10:44:01.846655654 +0000 UTC m=+498.758663392" watchObservedRunningTime="2026-02-28 10:44:01.850579443 +0000 UTC m=+498.762587181" Feb 28 10:44:02 crc kubenswrapper[4972]: I0228 10:44:02.843809 4972 generic.go:334] "Generic (PLEG): container finished" podID="1fcdc4df-a299-4e3c-970b-47c4e01aa51e" containerID="9fe3be65c4e68afcb7b9dfa412fad3a1caf9de652a625e28d6a44609f7d5f336" exitCode=0 Feb 28 10:44:02 crc kubenswrapper[4972]: I0228 10:44:02.843892 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" event={"ID":"1fcdc4df-a299-4e3c-970b-47c4e01aa51e","Type":"ContainerDied","Data":"9fe3be65c4e68afcb7b9dfa412fad3a1caf9de652a625e28d6a44609f7d5f336"} Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.068705 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.134222 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4cmx\" (UniqueName: \"kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx\") pod \"1fcdc4df-a299-4e3c-970b-47c4e01aa51e\" (UID: \"1fcdc4df-a299-4e3c-970b-47c4e01aa51e\") " Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.141859 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx" (OuterVolumeSpecName: "kube-api-access-m4cmx") pod "1fcdc4df-a299-4e3c-970b-47c4e01aa51e" (UID: "1fcdc4df-a299-4e3c-970b-47c4e01aa51e"). InnerVolumeSpecName "kube-api-access-m4cmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.235771 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4cmx\" (UniqueName: \"kubernetes.io/projected/1fcdc4df-a299-4e3c-970b-47c4e01aa51e-kube-api-access-m4cmx\") on node \"crc\" DevicePath \"\"" Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.859756 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" event={"ID":"1fcdc4df-a299-4e3c-970b-47c4e01aa51e","Type":"ContainerDied","Data":"c53cd232b006f58c9e1379f143519eb4cf288e867374d44ae4bbab471645de75"} Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.859839 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c53cd232b006f58c9e1379f143519eb4cf288e867374d44ae4bbab471645de75" Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.859855 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537924-h2gm7" Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.913677 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537918-nvtxp"] Feb 28 10:44:04 crc kubenswrapper[4972]: I0228 10:44:04.917529 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537918-nvtxp"] Feb 28 10:44:05 crc kubenswrapper[4972]: I0228 10:44:05.802120 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f27dc9-eb59-4aef-92e5-1267b61d4fad" path="/var/lib/kubelet/pods/36f27dc9-eb59-4aef-92e5-1267b61d4fad/volumes" Feb 28 10:44:28 crc kubenswrapper[4972]: I0228 10:44:28.890753 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:44:28 crc kubenswrapper[4972]: I0228 10:44:28.891607 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:44:28 crc kubenswrapper[4972]: I0228 10:44:28.891688 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:44:28 crc kubenswrapper[4972]: I0228 10:44:28.892707 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:44:28 crc kubenswrapper[4972]: I0228 10:44:28.892836 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd" gracePeriod=600 Feb 28 10:44:29 crc kubenswrapper[4972]: I0228 10:44:29.045612 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd" exitCode=0 Feb 28 10:44:29 crc kubenswrapper[4972]: I0228 10:44:29.046069 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd"} Feb 28 10:44:29 crc kubenswrapper[4972]: I0228 10:44:29.046326 4972 scope.go:117] "RemoveContainer" containerID="a7f4fb5eb4ee541f4e81f214080f816f771fdc1312a3799f99224983a7ebd094" Feb 28 10:44:30 crc kubenswrapper[4972]: I0228 10:44:30.054936 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298"} Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.148706 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6"] Feb 28 10:45:00 crc kubenswrapper[4972]: E0228 10:45:00.149617 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcdc4df-a299-4e3c-970b-47c4e01aa51e" containerName="oc" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.149632 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcdc4df-a299-4e3c-970b-47c4e01aa51e" containerName="oc" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.149731 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fcdc4df-a299-4e3c-970b-47c4e01aa51e" containerName="oc" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.150082 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.152492 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.155873 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.167826 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6"] Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.287405 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.287834 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.287875 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk2vw\" (UniqueName: \"kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.389808 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.389879 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.389943 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk2vw\" (UniqueName: \"kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.391071 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.401623 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.409020 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk2vw\" (UniqueName: \"kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw\") pod \"collect-profiles-29537925-tgtt6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.475608 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:00 crc kubenswrapper[4972]: I0228 10:45:00.727142 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6"] Feb 28 10:45:01 crc kubenswrapper[4972]: I0228 10:45:01.282726 4972 generic.go:334] "Generic (PLEG): container finished" podID="c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" containerID="471517a922347f1134b014ba802be3250743788a53d43efe6bbc797d79162ee7" exitCode=0 Feb 28 10:45:01 crc kubenswrapper[4972]: I0228 10:45:01.282781 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" event={"ID":"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6","Type":"ContainerDied","Data":"471517a922347f1134b014ba802be3250743788a53d43efe6bbc797d79162ee7"} Feb 28 10:45:01 crc kubenswrapper[4972]: I0228 10:45:01.282829 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" event={"ID":"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6","Type":"ContainerStarted","Data":"110b0ab8b4a36f0a7d7468d10a80376efc33c175e58e5caeed448d38b08c5c66"} Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.564375 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.647644 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume\") pod \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.647705 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk2vw\" (UniqueName: \"kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw\") pod \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.647779 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume\") pod \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\" (UID: \"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6\") " Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.649038 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume" (OuterVolumeSpecName: "config-volume") pod "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" (UID: "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.657646 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" (UID: "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.657855 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw" (OuterVolumeSpecName: "kube-api-access-lk2vw") pod "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" (UID: "c8dd23ca-32b0-46f1-a8c7-bb08c70727f6"). InnerVolumeSpecName "kube-api-access-lk2vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.749436 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.749550 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 10:45:02 crc kubenswrapper[4972]: I0228 10:45:02.749567 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk2vw\" (UniqueName: \"kubernetes.io/projected/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6-kube-api-access-lk2vw\") on node \"crc\" DevicePath \"\"" Feb 28 10:45:03 crc kubenswrapper[4972]: I0228 10:45:03.297770 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" event={"ID":"c8dd23ca-32b0-46f1-a8c7-bb08c70727f6","Type":"ContainerDied","Data":"110b0ab8b4a36f0a7d7468d10a80376efc33c175e58e5caeed448d38b08c5c66"} Feb 28 10:45:03 crc kubenswrapper[4972]: I0228 10:45:03.297836 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="110b0ab8b4a36f0a7d7468d10a80376efc33c175e58e5caeed448d38b08c5c66" Feb 28 10:45:03 crc kubenswrapper[4972]: I0228 10:45:03.297843 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6" Feb 28 10:45:56 crc kubenswrapper[4972]: I0228 10:45:56.896956 4972 scope.go:117] "RemoveContainer" containerID="a87faa3ec6ec59670c6dc44deed17678a472cdc8b758a80f31a7d5dfcce24ec4" Feb 28 10:45:56 crc kubenswrapper[4972]: I0228 10:45:56.926118 4972 scope.go:117] "RemoveContainer" containerID="06e657f11c83ddaaeeb67cf72e76f6b437469dad1f9f3897258650ef8527ef82" Feb 28 10:45:56 crc kubenswrapper[4972]: I0228 10:45:56.952580 4972 scope.go:117] "RemoveContainer" containerID="0bab11e9a64188cb3222a15eee1b23cb1acaf2d2bacf6bd87281b853e5ba246f" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.145671 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537926-krdlc"] Feb 28 10:46:00 crc kubenswrapper[4972]: E0228 10:46:00.146318 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" containerName="collect-profiles" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.146331 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" containerName="collect-profiles" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.146422 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" containerName="collect-profiles" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.146953 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.149249 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.149527 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.149737 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.167321 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537926-krdlc"] Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.333807 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rknkz\" (UniqueName: \"kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz\") pod \"auto-csr-approver-29537926-krdlc\" (UID: \"953c17df-68d7-407d-b186-17551b7b97b2\") " pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.435023 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rknkz\" (UniqueName: \"kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz\") pod \"auto-csr-approver-29537926-krdlc\" (UID: \"953c17df-68d7-407d-b186-17551b7b97b2\") " pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.462971 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rknkz\" (UniqueName: \"kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz\") pod \"auto-csr-approver-29537926-krdlc\" (UID: \"953c17df-68d7-407d-b186-17551b7b97b2\") " pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.476335 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:00 crc kubenswrapper[4972]: I0228 10:46:00.723696 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537926-krdlc"] Feb 28 10:46:01 crc kubenswrapper[4972]: I0228 10:46:01.710807 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537926-krdlc" event={"ID":"953c17df-68d7-407d-b186-17551b7b97b2","Type":"ContainerStarted","Data":"04ef5b7d52b0259f73cb034f3ff1f41316d32fdd3d7523f286cff6e85f60c637"} Feb 28 10:46:02 crc kubenswrapper[4972]: I0228 10:46:02.722516 4972 generic.go:334] "Generic (PLEG): container finished" podID="953c17df-68d7-407d-b186-17551b7b97b2" containerID="f71e6577840dbe4ba1155be8097374581a4a92b4f8ad1f8585f5b5befce71400" exitCode=0 Feb 28 10:46:02 crc kubenswrapper[4972]: I0228 10:46:02.722674 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537926-krdlc" event={"ID":"953c17df-68d7-407d-b186-17551b7b97b2","Type":"ContainerDied","Data":"f71e6577840dbe4ba1155be8097374581a4a92b4f8ad1f8585f5b5befce71400"} Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.018061 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.106255 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rknkz\" (UniqueName: \"kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz\") pod \"953c17df-68d7-407d-b186-17551b7b97b2\" (UID: \"953c17df-68d7-407d-b186-17551b7b97b2\") " Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.116623 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz" (OuterVolumeSpecName: "kube-api-access-rknkz") pod "953c17df-68d7-407d-b186-17551b7b97b2" (UID: "953c17df-68d7-407d-b186-17551b7b97b2"). InnerVolumeSpecName "kube-api-access-rknkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.207965 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rknkz\" (UniqueName: \"kubernetes.io/projected/953c17df-68d7-407d-b186-17551b7b97b2-kube-api-access-rknkz\") on node \"crc\" DevicePath \"\"" Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.741300 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537926-krdlc" event={"ID":"953c17df-68d7-407d-b186-17551b7b97b2","Type":"ContainerDied","Data":"04ef5b7d52b0259f73cb034f3ff1f41316d32fdd3d7523f286cff6e85f60c637"} Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.741371 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04ef5b7d52b0259f73cb034f3ff1f41316d32fdd3d7523f286cff6e85f60c637" Feb 28 10:46:04 crc kubenswrapper[4972]: I0228 10:46:04.741838 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537926-krdlc" Feb 28 10:46:05 crc kubenswrapper[4972]: I0228 10:46:05.079956 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537920-mpsj6"] Feb 28 10:46:05 crc kubenswrapper[4972]: I0228 10:46:05.088329 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537920-mpsj6"] Feb 28 10:46:05 crc kubenswrapper[4972]: I0228 10:46:05.801846 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e" path="/var/lib/kubelet/pods/4eccbb48-a68d-4b6c-b5cd-e4eddb7d1b7e/volumes" Feb 28 10:46:57 crc kubenswrapper[4972]: I0228 10:46:57.016001 4972 scope.go:117] "RemoveContainer" containerID="6a85130a8dd8f8b8faf3e7d1949e41c9bb84368575780ddb4d5df8bf0e86c615" Feb 28 10:46:58 crc kubenswrapper[4972]: I0228 10:46:58.890759 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:46:58 crc kubenswrapper[4972]: I0228 10:46:58.890861 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.248073 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h"] Feb 28 10:47:18 crc kubenswrapper[4972]: E0228 10:47:18.248835 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953c17df-68d7-407d-b186-17551b7b97b2" containerName="oc" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.248851 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="953c17df-68d7-407d-b186-17551b7b97b2" containerName="oc" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.248995 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="953c17df-68d7-407d-b186-17551b7b97b2" containerName="oc" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.249422 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.251618 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.252128 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.252378 4972 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-sjsxp" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.266733 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.278312 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-mrgf8"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.279210 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mrgf8" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.285210 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-76jg5"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.285900 4972 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-x6mbk" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.286251 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.288302 4972 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jwwkh" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.308135 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mrgf8"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.312439 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-76jg5"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.395330 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfb9\" (UniqueName: \"kubernetes.io/projected/c0da0fed-5127-4b89-a79f-73e3d9c1ceee-kube-api-access-hvfb9\") pod \"cert-manager-webhook-687f57d79b-76jg5\" (UID: \"c0da0fed-5127-4b89-a79f-73e3d9c1ceee\") " pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.395426 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtpt5\" (UniqueName: \"kubernetes.io/projected/74cd6520-9a2a-4e93-8536-2817d087ba65-kube-api-access-jtpt5\") pod \"cert-manager-858654f9db-mrgf8\" (UID: \"74cd6520-9a2a-4e93-8536-2817d087ba65\") " pod="cert-manager/cert-manager-858654f9db-mrgf8" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.395545 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbrxb\" (UniqueName: \"kubernetes.io/projected/0dcc4f20-2ac3-4a75-8674-dfe9800d1844-kube-api-access-wbrxb\") pod \"cert-manager-cainjector-cf98fcc89-ptx2h\" (UID: \"0dcc4f20-2ac3-4a75-8674-dfe9800d1844\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.497045 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbrxb\" (UniqueName: \"kubernetes.io/projected/0dcc4f20-2ac3-4a75-8674-dfe9800d1844-kube-api-access-wbrxb\") pod \"cert-manager-cainjector-cf98fcc89-ptx2h\" (UID: \"0dcc4f20-2ac3-4a75-8674-dfe9800d1844\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.497126 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfb9\" (UniqueName: \"kubernetes.io/projected/c0da0fed-5127-4b89-a79f-73e3d9c1ceee-kube-api-access-hvfb9\") pod \"cert-manager-webhook-687f57d79b-76jg5\" (UID: \"c0da0fed-5127-4b89-a79f-73e3d9c1ceee\") " pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.497197 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtpt5\" (UniqueName: \"kubernetes.io/projected/74cd6520-9a2a-4e93-8536-2817d087ba65-kube-api-access-jtpt5\") pod \"cert-manager-858654f9db-mrgf8\" (UID: \"74cd6520-9a2a-4e93-8536-2817d087ba65\") " pod="cert-manager/cert-manager-858654f9db-mrgf8" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.522840 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbrxb\" (UniqueName: \"kubernetes.io/projected/0dcc4f20-2ac3-4a75-8674-dfe9800d1844-kube-api-access-wbrxb\") pod \"cert-manager-cainjector-cf98fcc89-ptx2h\" (UID: \"0dcc4f20-2ac3-4a75-8674-dfe9800d1844\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.523226 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfb9\" (UniqueName: \"kubernetes.io/projected/c0da0fed-5127-4b89-a79f-73e3d9c1ceee-kube-api-access-hvfb9\") pod \"cert-manager-webhook-687f57d79b-76jg5\" (UID: \"c0da0fed-5127-4b89-a79f-73e3d9c1ceee\") " pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.526790 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtpt5\" (UniqueName: \"kubernetes.io/projected/74cd6520-9a2a-4e93-8536-2817d087ba65-kube-api-access-jtpt5\") pod \"cert-manager-858654f9db-mrgf8\" (UID: \"74cd6520-9a2a-4e93-8536-2817d087ba65\") " pod="cert-manager/cert-manager-858654f9db-mrgf8" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.569425 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.597636 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-mrgf8" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.606093 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.825751 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h"] Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.858012 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-76jg5"] Feb 28 10:47:18 crc kubenswrapper[4972]: W0228 10:47:18.868885 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0da0fed_5127_4b89_a79f_73e3d9c1ceee.slice/crio-cfb1a46219e0b6ee9ff1e9b2c22b2c615bd24ce99bb9f1f67815b7500d210668 WatchSource:0}: Error finding container cfb1a46219e0b6ee9ff1e9b2c22b2c615bd24ce99bb9f1f67815b7500d210668: Status 404 returned error can't find the container with id cfb1a46219e0b6ee9ff1e9b2c22b2c615bd24ce99bb9f1f67815b7500d210668 Feb 28 10:47:18 crc kubenswrapper[4972]: I0228 10:47:18.912925 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-mrgf8"] Feb 28 10:47:18 crc kubenswrapper[4972]: W0228 10:47:18.918329 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74cd6520_9a2a_4e93_8536_2817d087ba65.slice/crio-fb24852ea765cf7c9f67bb710c3df2fa5e3e173102cadac86b264affff0b99e1 WatchSource:0}: Error finding container fb24852ea765cf7c9f67bb710c3df2fa5e3e173102cadac86b264affff0b99e1: Status 404 returned error can't find the container with id fb24852ea765cf7c9f67bb710c3df2fa5e3e173102cadac86b264affff0b99e1 Feb 28 10:47:19 crc kubenswrapper[4972]: I0228 10:47:19.263700 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" event={"ID":"c0da0fed-5127-4b89-a79f-73e3d9c1ceee","Type":"ContainerStarted","Data":"cfb1a46219e0b6ee9ff1e9b2c22b2c615bd24ce99bb9f1f67815b7500d210668"} Feb 28 10:47:19 crc kubenswrapper[4972]: I0228 10:47:19.265289 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mrgf8" event={"ID":"74cd6520-9a2a-4e93-8536-2817d087ba65","Type":"ContainerStarted","Data":"fb24852ea765cf7c9f67bb710c3df2fa5e3e173102cadac86b264affff0b99e1"} Feb 28 10:47:19 crc kubenswrapper[4972]: I0228 10:47:19.267054 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" event={"ID":"0dcc4f20-2ac3-4a75-8674-dfe9800d1844","Type":"ContainerStarted","Data":"e48fb10f30198604b495c96478df3152414f56e47e1958b34ea0e66d3dc1f947"} Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.290111 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" event={"ID":"c0da0fed-5127-4b89-a79f-73e3d9c1ceee","Type":"ContainerStarted","Data":"b3213632f07c4e7ed2bf8729c0403d7c4a7fc83a554274a5790621d1938eaf22"} Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.291034 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.292434 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-mrgf8" event={"ID":"74cd6520-9a2a-4e93-8536-2817d087ba65","Type":"ContainerStarted","Data":"e143b8fb436afc99fd603e4b07b6ee2a70f5ae25ba44aff1bf947bbb8d8fcf6a"} Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.294072 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" event={"ID":"0dcc4f20-2ac3-4a75-8674-dfe9800d1844","Type":"ContainerStarted","Data":"9ccc7a3c28fba8cbe7a22e1f544789822badacc1bed7cfde6e73830f1ad339d7"} Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.311279 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" podStartSLOduration=1.232271439 podStartE2EDuration="5.311251932s" podCreationTimestamp="2026-02-28 10:47:18 +0000 UTC" firstStartedPulling="2026-02-28 10:47:18.873343725 +0000 UTC m=+695.785351463" lastFinishedPulling="2026-02-28 10:47:22.952324188 +0000 UTC m=+699.864331956" observedRunningTime="2026-02-28 10:47:23.308216237 +0000 UTC m=+700.220223965" watchObservedRunningTime="2026-02-28 10:47:23.311251932 +0000 UTC m=+700.223259670" Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.328413 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-ptx2h" podStartSLOduration=1.275558903 podStartE2EDuration="5.328383292s" podCreationTimestamp="2026-02-28 10:47:18 +0000 UTC" firstStartedPulling="2026-02-28 10:47:18.83257415 +0000 UTC m=+695.744581888" lastFinishedPulling="2026-02-28 10:47:22.885398539 +0000 UTC m=+699.797406277" observedRunningTime="2026-02-28 10:47:23.327822047 +0000 UTC m=+700.239829785" watchObservedRunningTime="2026-02-28 10:47:23.328383292 +0000 UTC m=+700.240391030" Feb 28 10:47:23 crc kubenswrapper[4972]: I0228 10:47:23.345581 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-mrgf8" podStartSLOduration=1.379973384 podStartE2EDuration="5.345553694s" podCreationTimestamp="2026-02-28 10:47:18 +0000 UTC" firstStartedPulling="2026-02-28 10:47:18.920661593 +0000 UTC m=+695.832669321" lastFinishedPulling="2026-02-28 10:47:22.886241853 +0000 UTC m=+699.798249631" observedRunningTime="2026-02-28 10:47:23.343921499 +0000 UTC m=+700.255929237" watchObservedRunningTime="2026-02-28 10:47:23.345553694 +0000 UTC m=+700.257561432" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.384445 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-crbqv"] Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385176 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-controller" containerID="cri-o://9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385294 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385277 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="sbdb" containerID="cri-o://87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385325 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-node" containerID="cri-o://dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385327 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="northd" containerID="cri-o://ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385527 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" containerID="cri-o://0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.385567 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="nbdb" containerID="cri-o://db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.435496 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovnkube-controller" containerID="cri-o://41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" gracePeriod=30 Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.611526 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-76jg5" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.742993 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/1.log" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.747196 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.747824 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-controller/0.log" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.748352 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.817428 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gt8hp"] Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.817874 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="nbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.817949 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="nbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818002 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818052 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818109 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-ovn-metrics" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818160 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-ovn-metrics" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818209 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818258 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818310 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kubecfg-setup" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818362 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kubecfg-setup" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818413 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818505 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818618 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-node" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818679 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-node" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818764 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="northd" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818816 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="northd" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818882 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="sbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.818937 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="sbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: E0228 10:47:28.818997 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovnkube-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819043 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovnkube-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819215 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-node" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819279 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="kube-rbac-proxy-ovn-metrics" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819328 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="northd" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819391 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819444 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819518 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="sbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819578 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovnkube-controller" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819630 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="nbdb" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.819684 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerName="ovn-acl-logging" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.821522 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.850868 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.851081 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.851235 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.852150 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.852276 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.852373 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854665 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854744 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854797 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854858 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fknn\" (UniqueName: \"kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854936 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855074 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855127 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855176 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855278 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855362 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855416 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855495 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855570 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855651 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch\") pod \"4f815683-30aa-4ea4-928d-6ba5e7945de0\" (UID: \"4f815683-30aa-4ea4-928d-6ba5e7945de0\") " Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.851145 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash" (OuterVolumeSpecName: "host-slash") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.851601 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.851756 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.852242 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.854026 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.855963 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856047 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket" (OuterVolumeSpecName: "log-socket") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856100 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856514 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856571 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log" (OuterVolumeSpecName: "node-log") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856614 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856669 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.856723 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857026 4972 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857062 4972 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857084 4972 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-log-socket\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857102 4972 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857118 4972 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-node-log\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857134 4972 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857151 4972 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857168 4972 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857184 4972 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857201 4972 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-slash\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857220 4972 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857237 4972 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857253 4972 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857247 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857311 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857554 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.857723 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.862267 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn" (OuterVolumeSpecName: "kube-api-access-6fknn") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "kube-api-access-6fknn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.863733 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.869581 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4f815683-30aa-4ea4-928d-6ba5e7945de0" (UID: "4f815683-30aa-4ea4-928d-6ba5e7945de0"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.891154 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.891217 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958019 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-bin\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958078 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958116 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrld7\" (UniqueName: \"kubernetes.io/projected/362372a4-96fb-4bef-8f82-42acd4970edc-kube-api-access-hrld7\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958154 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-etc-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958196 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-systemd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958221 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-var-lib-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958251 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-node-log\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958307 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-slash\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958339 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-env-overrides\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958445 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-script-lib\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958501 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-ovn\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958531 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-netns\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958561 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/362372a4-96fb-4bef-8f82-42acd4970edc-ovn-node-metrics-cert\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958600 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-netd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958623 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-systemd-units\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958638 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-config\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958655 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958673 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958733 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-kubelet\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958759 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-log-socket\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958794 4972 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958805 4972 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958814 4972 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958823 4972 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958831 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fknn\" (UniqueName: \"kubernetes.io/projected/4f815683-30aa-4ea4-928d-6ba5e7945de0-kube-api-access-6fknn\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958839 4972 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f815683-30aa-4ea4-928d-6ba5e7945de0-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:28 crc kubenswrapper[4972]: I0228 10:47:28.958847 4972 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f815683-30aa-4ea4-928d-6ba5e7945de0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.059907 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-script-lib\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.059974 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-ovn\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060019 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-netns\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060047 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/362372a4-96fb-4bef-8f82-42acd4970edc-ovn-node-metrics-cert\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-netd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060100 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-systemd-units\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-config\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060146 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-ovn\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060155 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060230 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060283 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-kubelet\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060298 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-run-netns\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060365 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-log-socket\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060376 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060424 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-kubelet\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060303 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-systemd-units\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060237 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-netd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060326 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-log-socket\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060592 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-bin\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060631 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060668 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-cni-bin\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060682 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrld7\" (UniqueName: \"kubernetes.io/projected/362372a4-96fb-4bef-8f82-42acd4970edc-kube-api-access-hrld7\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060749 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-etc-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060819 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060826 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-systemd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060852 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-run-systemd\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060886 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-etc-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060886 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-var-lib-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060922 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-node-log\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060932 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-var-lib-openvswitch\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.060982 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-node-log\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061004 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-slash\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061088 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/362372a4-96fb-4bef-8f82-42acd4970edc-host-slash\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061137 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-env-overrides\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061182 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-config\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061190 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-ovnkube-script-lib\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.061583 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/362372a4-96fb-4bef-8f82-42acd4970edc-env-overrides\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.063570 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/362372a4-96fb-4bef-8f82-42acd4970edc-ovn-node-metrics-cert\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.080968 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrld7\" (UniqueName: \"kubernetes.io/projected/362372a4-96fb-4bef-8f82-42acd4970edc-kube-api-access-hrld7\") pod \"ovnkube-node-gt8hp\" (UID: \"362372a4-96fb-4bef-8f82-42acd4970edc\") " pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.135762 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.333815 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pl4fx_0302d7e4-875d-479b-9773-683e7f17a9f7/kube-multus/0.log" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.334297 4972 generic.go:334] "Generic (PLEG): container finished" podID="0302d7e4-875d-479b-9773-683e7f17a9f7" containerID="fdb2cc50d96375b8f53e4e976b0249d68f023875839c7bf59a00755f3c0209fe" exitCode=2 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.334356 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pl4fx" event={"ID":"0302d7e4-875d-479b-9773-683e7f17a9f7","Type":"ContainerDied","Data":"fdb2cc50d96375b8f53e4e976b0249d68f023875839c7bf59a00755f3c0209fe"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.334866 4972 scope.go:117] "RemoveContainer" containerID="fdb2cc50d96375b8f53e4e976b0249d68f023875839c7bf59a00755f3c0209fe" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.339240 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/1.log" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.344103 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-acl-logging/0.log" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.358159 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-crbqv_4f815683-30aa-4ea4-928d-6ba5e7945de0/ovn-controller/0.log" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.360905 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" exitCode=143 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361106 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361191 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361102 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361348 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361265 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361406 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361409 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361424 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361439 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361447 4972 generic.go:334] "Generic (PLEG): container finished" podID="4f815683-30aa-4ea4-928d-6ba5e7945de0" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" exitCode=143 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361142 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361377 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361565 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361589 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361602 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361612 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361621 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361626 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361631 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361636 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361641 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361646 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361652 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361657 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361663 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361673 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361680 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361686 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361692 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361698 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361703 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361708 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361713 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361717 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361722 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361729 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361736 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361741 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361749 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361755 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361772 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361779 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361786 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361792 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361798 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361804 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361837 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-crbqv" event={"ID":"4f815683-30aa-4ea4-928d-6ba5e7945de0","Type":"ContainerDied","Data":"e1546c6f07613d1f40ca78c524eedebc0e56d784610aeebb253389434f007566"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361849 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361857 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361863 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361869 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361875 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361880 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361887 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361891 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361897 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.361904 4972 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.363891 4972 generic.go:334] "Generic (PLEG): container finished" podID="362372a4-96fb-4bef-8f82-42acd4970edc" containerID="63c2796d90ab33684801ee77de1910cd0a88f7686649778509c9a01f6de0a9f3" exitCode=0 Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.363913 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerDied","Data":"63c2796d90ab33684801ee77de1910cd0a88f7686649778509c9a01f6de0a9f3"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.363928 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"1c6a89947068f73bc0f0dea010158f5c1b7ed6830cc4f8d3a5687617290c1025"} Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.410126 4972 scope.go:117] "RemoveContainer" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.427409 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-crbqv"] Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.431046 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-crbqv"] Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.460121 4972 scope.go:117] "RemoveContainer" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.475845 4972 scope.go:117] "RemoveContainer" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.488978 4972 scope.go:117] "RemoveContainer" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.502145 4972 scope.go:117] "RemoveContainer" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.517263 4972 scope.go:117] "RemoveContainer" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.584016 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.616672 4972 scope.go:117] "RemoveContainer" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.651584 4972 scope.go:117] "RemoveContainer" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.667278 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.667798 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.667841 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} err="failed to get container status \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.667869 4972 scope.go:117] "RemoveContainer" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.668647 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": container with ID starting with 41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca not found: ID does not exist" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.668704 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} err="failed to get container status \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": rpc error: code = NotFound desc = could not find container \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": container with ID starting with 41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.668745 4972 scope.go:117] "RemoveContainer" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.669112 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": container with ID starting with 87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a not found: ID does not exist" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.669139 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} err="failed to get container status \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": rpc error: code = NotFound desc = could not find container \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": container with ID starting with 87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.669156 4972 scope.go:117] "RemoveContainer" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.669593 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": container with ID starting with db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed not found: ID does not exist" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.669654 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} err="failed to get container status \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": rpc error: code = NotFound desc = could not find container \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": container with ID starting with db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.669692 4972 scope.go:117] "RemoveContainer" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.670037 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": container with ID starting with ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47 not found: ID does not exist" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.670082 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} err="failed to get container status \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": rpc error: code = NotFound desc = could not find container \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": container with ID starting with ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.670119 4972 scope.go:117] "RemoveContainer" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.670686 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": container with ID starting with d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea not found: ID does not exist" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.670712 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} err="failed to get container status \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": rpc error: code = NotFound desc = could not find container \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": container with ID starting with d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.670726 4972 scope.go:117] "RemoveContainer" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.671311 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": container with ID starting with dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb not found: ID does not exist" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.671398 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} err="failed to get container status \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": rpc error: code = NotFound desc = could not find container \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": container with ID starting with dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.671432 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.671851 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": container with ID starting with 97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788 not found: ID does not exist" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.671881 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} err="failed to get container status \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": rpc error: code = NotFound desc = could not find container \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": container with ID starting with 97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.671903 4972 scope.go:117] "RemoveContainer" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.672288 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": container with ID starting with 9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc not found: ID does not exist" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.672324 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} err="failed to get container status \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": rpc error: code = NotFound desc = could not find container \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": container with ID starting with 9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.672351 4972 scope.go:117] "RemoveContainer" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: E0228 10:47:29.672770 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": container with ID starting with f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0 not found: ID does not exist" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.672811 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} err="failed to get container status \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": rpc error: code = NotFound desc = could not find container \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": container with ID starting with f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.672838 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.673926 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} err="failed to get container status \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.673968 4972 scope.go:117] "RemoveContainer" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.674336 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} err="failed to get container status \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": rpc error: code = NotFound desc = could not find container \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": container with ID starting with 41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.674371 4972 scope.go:117] "RemoveContainer" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.674697 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} err="failed to get container status \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": rpc error: code = NotFound desc = could not find container \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": container with ID starting with 87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.674729 4972 scope.go:117] "RemoveContainer" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.676997 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} err="failed to get container status \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": rpc error: code = NotFound desc = could not find container \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": container with ID starting with db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.677035 4972 scope.go:117] "RemoveContainer" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.677633 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} err="failed to get container status \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": rpc error: code = NotFound desc = could not find container \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": container with ID starting with ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.677669 4972 scope.go:117] "RemoveContainer" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.679335 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} err="failed to get container status \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": rpc error: code = NotFound desc = could not find container \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": container with ID starting with d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.679365 4972 scope.go:117] "RemoveContainer" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.679817 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} err="failed to get container status \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": rpc error: code = NotFound desc = could not find container \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": container with ID starting with dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.679860 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.680139 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} err="failed to get container status \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": rpc error: code = NotFound desc = could not find container \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": container with ID starting with 97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.680165 4972 scope.go:117] "RemoveContainer" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.680516 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} err="failed to get container status \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": rpc error: code = NotFound desc = could not find container \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": container with ID starting with 9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.680595 4972 scope.go:117] "RemoveContainer" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.681134 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} err="failed to get container status \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": rpc error: code = NotFound desc = could not find container \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": container with ID starting with f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.681162 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.681616 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} err="failed to get container status \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.681645 4972 scope.go:117] "RemoveContainer" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682072 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} err="failed to get container status \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": rpc error: code = NotFound desc = could not find container \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": container with ID starting with 41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682098 4972 scope.go:117] "RemoveContainer" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682479 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} err="failed to get container status \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": rpc error: code = NotFound desc = could not find container \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": container with ID starting with 87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682505 4972 scope.go:117] "RemoveContainer" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682924 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} err="failed to get container status \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": rpc error: code = NotFound desc = could not find container \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": container with ID starting with db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.682949 4972 scope.go:117] "RemoveContainer" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.683645 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} err="failed to get container status \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": rpc error: code = NotFound desc = could not find container \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": container with ID starting with ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.683674 4972 scope.go:117] "RemoveContainer" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.683997 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} err="failed to get container status \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": rpc error: code = NotFound desc = could not find container \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": container with ID starting with d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.684034 4972 scope.go:117] "RemoveContainer" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.684557 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} err="failed to get container status \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": rpc error: code = NotFound desc = could not find container \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": container with ID starting with dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.684611 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.686622 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} err="failed to get container status \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": rpc error: code = NotFound desc = could not find container \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": container with ID starting with 97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.686663 4972 scope.go:117] "RemoveContainer" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.686962 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} err="failed to get container status \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": rpc error: code = NotFound desc = could not find container \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": container with ID starting with 9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.686990 4972 scope.go:117] "RemoveContainer" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.687448 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} err="failed to get container status \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": rpc error: code = NotFound desc = could not find container \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": container with ID starting with f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.687528 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688037 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} err="failed to get container status \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688078 4972 scope.go:117] "RemoveContainer" containerID="41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688400 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca"} err="failed to get container status \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": rpc error: code = NotFound desc = could not find container \"41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca\": container with ID starting with 41f7c04a926aa5e15bea5fd9d97c7471be9905bd407669d56125faffe8def7ca not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688429 4972 scope.go:117] "RemoveContainer" containerID="87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688765 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a"} err="failed to get container status \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": rpc error: code = NotFound desc = could not find container \"87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a\": container with ID starting with 87664e8b961d1b452d0ba6865982cb704bacb20af71541116a7aecde69d3aa1a not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.688814 4972 scope.go:117] "RemoveContainer" containerID="db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689209 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed"} err="failed to get container status \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": rpc error: code = NotFound desc = could not find container \"db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed\": container with ID starting with db551bf5438e90718a5786a8a44516bdd3742b414625201689dd97c8294b74ed not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689236 4972 scope.go:117] "RemoveContainer" containerID="ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689622 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47"} err="failed to get container status \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": rpc error: code = NotFound desc = could not find container \"ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47\": container with ID starting with ed01e8ebce3924648267feb566e698023d25efffede2f67446ed877ce17d4e47 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689662 4972 scope.go:117] "RemoveContainer" containerID="d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689949 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea"} err="failed to get container status \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": rpc error: code = NotFound desc = could not find container \"d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea\": container with ID starting with d74646f9ef6e34f99c461f9d6ee6287ca29537981b1a807b658e4868e83f00ea not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.689974 4972 scope.go:117] "RemoveContainer" containerID="dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.690288 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb"} err="failed to get container status \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": rpc error: code = NotFound desc = could not find container \"dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb\": container with ID starting with dd21cf501109796e0937921726d6e1c122663bb61c552c96729f9777b7d3bbdb not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.690330 4972 scope.go:117] "RemoveContainer" containerID="97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.690783 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788"} err="failed to get container status \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": rpc error: code = NotFound desc = could not find container \"97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788\": container with ID starting with 97473c8e89b728d5d8555d11944315c8c45e671489bef6bbd07ad07df3ca1788 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.690824 4972 scope.go:117] "RemoveContainer" containerID="9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.691171 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc"} err="failed to get container status \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": rpc error: code = NotFound desc = could not find container \"9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc\": container with ID starting with 9bd9d98d3139ae0e87cad90ac162e70bb799172da0bf48ee99045a4cfead2cdc not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.691230 4972 scope.go:117] "RemoveContainer" containerID="f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.691510 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0"} err="failed to get container status \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": rpc error: code = NotFound desc = could not find container \"f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0\": container with ID starting with f402c2b25f5fc0df95efb9870c150282a763eba357ea018facb2d0a1196761c0 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.691533 4972 scope.go:117] "RemoveContainer" containerID="0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.691824 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7"} err="failed to get container status \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": rpc error: code = NotFound desc = could not find container \"0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7\": container with ID starting with 0157007a69e6333868315d4e442d53d017751a246bf84f6a986c14bf2e97edd7 not found: ID does not exist" Feb 28 10:47:29 crc kubenswrapper[4972]: I0228 10:47:29.798972 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f815683-30aa-4ea4-928d-6ba5e7945de0" path="/var/lib/kubelet/pods/4f815683-30aa-4ea4-928d-6ba5e7945de0/volumes" Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.372501 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pl4fx_0302d7e4-875d-479b-9773-683e7f17a9f7/kube-multus/0.log" Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.372663 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pl4fx" event={"ID":"0302d7e4-875d-479b-9773-683e7f17a9f7","Type":"ContainerStarted","Data":"406468e38288a02783cd57aca217e1561eeee71721167596fe72d9074976833d"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379615 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"928877121da7e900527725ae06988b8913874e5b63ee5e78c3c187926d43e0d7"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379672 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"d6ad52387b4bc7c7bef89b4a9db24e859ec6390ac5bd37e81cf89ba92bbb7048"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379686 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"eb0e246ef7f872c5df0ecf8c8e634e3ffd123e6f9c0ddd429bcc4f1dab99b78c"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379738 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"073412981afbf815cd819f88a3d5ceac62796fe6c051d9ba01396ddff48d3115"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379750 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"9c0ee769f36472c496c0584a5ca99734adcb889d24a553d8cff04b09dd23508e"} Feb 28 10:47:30 crc kubenswrapper[4972]: I0228 10:47:30.379760 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"37d478701dd39bb201341c6fb307dc72a3e4ddda0c303f73b42dac6299996f98"} Feb 28 10:47:33 crc kubenswrapper[4972]: I0228 10:47:33.404001 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"f043acd087ca0faae2f9aadeb6c1d6acb69c335f1d1730da0226a5b0cea3d357"} Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.418996 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" event={"ID":"362372a4-96fb-4bef-8f82-42acd4970edc","Type":"ContainerStarted","Data":"9a2ed2e61b1c254b54414d08367fc051b003fe3a8ef052c076e02bb2ab0b5aed"} Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.419831 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.419969 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.419981 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.448265 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.451512 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:35 crc kubenswrapper[4972]: I0228 10:47:35.496713 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" podStartSLOduration=7.496693896 podStartE2EDuration="7.496693896s" podCreationTimestamp="2026-02-28 10:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:47:35.464953185 +0000 UTC m=+712.376960973" watchObservedRunningTime="2026-02-28 10:47:35.496693896 +0000 UTC m=+712.408701634" Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.487788 4972 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.890273 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.890701 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.890758 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.891405 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:47:58 crc kubenswrapper[4972]: I0228 10:47:58.891496 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298" gracePeriod=600 Feb 28 10:47:59 crc kubenswrapper[4972]: I0228 10:47:59.166053 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gt8hp" Feb 28 10:47:59 crc kubenswrapper[4972]: I0228 10:47:59.617780 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298" exitCode=0 Feb 28 10:47:59 crc kubenswrapper[4972]: I0228 10:47:59.618219 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298"} Feb 28 10:47:59 crc kubenswrapper[4972]: I0228 10:47:59.618255 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3"} Feb 28 10:47:59 crc kubenswrapper[4972]: I0228 10:47:59.618276 4972 scope.go:117] "RemoveContainer" containerID="c671d8fc00a4b2e0641c3cd9619a400d4bfd5f3ed0ee9f9485f1263b047851cd" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.154223 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537928-lpsh8"] Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.155596 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.159181 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.173211 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.181138 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.199679 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537928-lpsh8"] Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.339935 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f6mv\" (UniqueName: \"kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv\") pod \"auto-csr-approver-29537928-lpsh8\" (UID: \"96ffb7ef-d239-4a03-8654-641bcad251bd\") " pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.441454 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f6mv\" (UniqueName: \"kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv\") pod \"auto-csr-approver-29537928-lpsh8\" (UID: \"96ffb7ef-d239-4a03-8654-641bcad251bd\") " pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.471841 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f6mv\" (UniqueName: \"kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv\") pod \"auto-csr-approver-29537928-lpsh8\" (UID: \"96ffb7ef-d239-4a03-8654-641bcad251bd\") " pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.492251 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:00 crc kubenswrapper[4972]: I0228 10:48:00.720662 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537928-lpsh8"] Feb 28 10:48:01 crc kubenswrapper[4972]: I0228 10:48:01.641678 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" event={"ID":"96ffb7ef-d239-4a03-8654-641bcad251bd","Type":"ContainerStarted","Data":"cf0c971e8ffa876052ee6ce9e42715030bd112f01564c87cab2d337058a47781"} Feb 28 10:48:02 crc kubenswrapper[4972]: I0228 10:48:02.650451 4972 generic.go:334] "Generic (PLEG): container finished" podID="96ffb7ef-d239-4a03-8654-641bcad251bd" containerID="625df873be954ef27d6136bf8de206a89bf44bfaec7dd2f36cd25ca732df513a" exitCode=0 Feb 28 10:48:02 crc kubenswrapper[4972]: I0228 10:48:02.650644 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" event={"ID":"96ffb7ef-d239-4a03-8654-641bcad251bd","Type":"ContainerDied","Data":"625df873be954ef27d6136bf8de206a89bf44bfaec7dd2f36cd25ca732df513a"} Feb 28 10:48:03 crc kubenswrapper[4972]: I0228 10:48:03.937872 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.006127 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f6mv\" (UniqueName: \"kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv\") pod \"96ffb7ef-d239-4a03-8654-641bcad251bd\" (UID: \"96ffb7ef-d239-4a03-8654-641bcad251bd\") " Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.013210 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv" (OuterVolumeSpecName: "kube-api-access-2f6mv") pod "96ffb7ef-d239-4a03-8654-641bcad251bd" (UID: "96ffb7ef-d239-4a03-8654-641bcad251bd"). InnerVolumeSpecName "kube-api-access-2f6mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.107847 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f6mv\" (UniqueName: \"kubernetes.io/projected/96ffb7ef-d239-4a03-8654-641bcad251bd-kube-api-access-2f6mv\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.668581 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" event={"ID":"96ffb7ef-d239-4a03-8654-641bcad251bd","Type":"ContainerDied","Data":"cf0c971e8ffa876052ee6ce9e42715030bd112f01564c87cab2d337058a47781"} Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.668646 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf0c971e8ffa876052ee6ce9e42715030bd112f01564c87cab2d337058a47781" Feb 28 10:48:04 crc kubenswrapper[4972]: I0228 10:48:04.668698 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537928-lpsh8" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.005668 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537922-jstkp"] Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.010486 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537922-jstkp"] Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.800518 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c246d4-f5f0-4d59-8f40-3d891edb2cff" path="/var/lib/kubelet/pods/12c246d4-f5f0-4d59-8f40-3d891edb2cff/volumes" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.860300 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm"] Feb 28 10:48:05 crc kubenswrapper[4972]: E0228 10:48:05.860851 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ffb7ef-d239-4a03-8654-641bcad251bd" containerName="oc" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.860943 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ffb7ef-d239-4a03-8654-641bcad251bd" containerName="oc" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.861135 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ffb7ef-d239-4a03-8654-641bcad251bd" containerName="oc" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.862201 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.864391 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 28 10:48:05 crc kubenswrapper[4972]: I0228 10:48:05.871529 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm"] Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.041908 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.042059 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.042861 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-668nm\" (UniqueName: \"kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.143879 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.143938 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.143980 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-668nm\" (UniqueName: \"kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.144833 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.145152 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.165251 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-668nm\" (UniqueName: \"kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.187597 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.520299 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm"] Feb 28 10:48:06 crc kubenswrapper[4972]: W0228 10:48:06.530622 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eabfdfb_f2b5_458a_8f35_8c1aff548e91.slice/crio-aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56 WatchSource:0}: Error finding container aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56: Status 404 returned error can't find the container with id aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56 Feb 28 10:48:06 crc kubenswrapper[4972]: I0228 10:48:06.683059 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" event={"ID":"2eabfdfb-f2b5-458a-8f35-8c1aff548e91","Type":"ContainerStarted","Data":"aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56"} Feb 28 10:48:07 crc kubenswrapper[4972]: I0228 10:48:07.695221 4972 generic.go:334] "Generic (PLEG): container finished" podID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerID="f2ddd09d225b32135c370d65cff4a63febeefe1d4bd0b2766eb5686a22f1b709" exitCode=0 Feb 28 10:48:07 crc kubenswrapper[4972]: I0228 10:48:07.695297 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" event={"ID":"2eabfdfb-f2b5-458a-8f35-8c1aff548e91","Type":"ContainerDied","Data":"f2ddd09d225b32135c370d65cff4a63febeefe1d4bd0b2766eb5686a22f1b709"} Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.001082 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.005723 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.024172 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.174260 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.174522 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.174571 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zhf7\" (UniqueName: \"kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.275634 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.275711 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zhf7\" (UniqueName: \"kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.275758 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.276419 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.276583 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.299603 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zhf7\" (UniqueName: \"kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7\") pod \"redhat-operators-txp88\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.338923 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:08 crc kubenswrapper[4972]: I0228 10:48:08.752742 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:09 crc kubenswrapper[4972]: I0228 10:48:09.710962 4972 generic.go:334] "Generic (PLEG): container finished" podID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerID="8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294" exitCode=0 Feb 28 10:48:09 crc kubenswrapper[4972]: I0228 10:48:09.711089 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerDied","Data":"8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294"} Feb 28 10:48:09 crc kubenswrapper[4972]: I0228 10:48:09.711755 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerStarted","Data":"f19b7f95b976af846e06725f815905ac283b89ceff7f0a30a0b482d879dc5941"} Feb 28 10:48:09 crc kubenswrapper[4972]: I0228 10:48:09.715149 4972 generic.go:334] "Generic (PLEG): container finished" podID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerID="29cbd44c2b76b4b79bb6daaf06d3792646137fad85dfc1fe95fd0de03efffb43" exitCode=0 Feb 28 10:48:09 crc kubenswrapper[4972]: I0228 10:48:09.715212 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" event={"ID":"2eabfdfb-f2b5-458a-8f35-8c1aff548e91","Type":"ContainerDied","Data":"29cbd44c2b76b4b79bb6daaf06d3792646137fad85dfc1fe95fd0de03efffb43"} Feb 28 10:48:10 crc kubenswrapper[4972]: I0228 10:48:10.727724 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerStarted","Data":"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb"} Feb 28 10:48:10 crc kubenswrapper[4972]: I0228 10:48:10.730991 4972 generic.go:334] "Generic (PLEG): container finished" podID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerID="45a522e8f66eaf85f72af5e7eddea0e8741756f543d4dff21328ea3d70cf2443" exitCode=0 Feb 28 10:48:10 crc kubenswrapper[4972]: I0228 10:48:10.731022 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" event={"ID":"2eabfdfb-f2b5-458a-8f35-8c1aff548e91","Type":"ContainerDied","Data":"45a522e8f66eaf85f72af5e7eddea0e8741756f543d4dff21328ea3d70cf2443"} Feb 28 10:48:11 crc kubenswrapper[4972]: I0228 10:48:11.740574 4972 generic.go:334] "Generic (PLEG): container finished" podID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerID="de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb" exitCode=0 Feb 28 10:48:11 crc kubenswrapper[4972]: I0228 10:48:11.740661 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerDied","Data":"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb"} Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.101731 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.231023 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle\") pod \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.231286 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-668nm\" (UniqueName: \"kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm\") pod \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.231553 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util\") pod \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\" (UID: \"2eabfdfb-f2b5-458a-8f35-8c1aff548e91\") " Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.233162 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle" (OuterVolumeSpecName: "bundle") pod "2eabfdfb-f2b5-458a-8f35-8c1aff548e91" (UID: "2eabfdfb-f2b5-458a-8f35-8c1aff548e91"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.234431 4972 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.241871 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm" (OuterVolumeSpecName: "kube-api-access-668nm") pod "2eabfdfb-f2b5-458a-8f35-8c1aff548e91" (UID: "2eabfdfb-f2b5-458a-8f35-8c1aff548e91"). InnerVolumeSpecName "kube-api-access-668nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.335524 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-668nm\" (UniqueName: \"kubernetes.io/projected/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-kube-api-access-668nm\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.339377 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util" (OuterVolumeSpecName: "util") pod "2eabfdfb-f2b5-458a-8f35-8c1aff548e91" (UID: "2eabfdfb-f2b5-458a-8f35-8c1aff548e91"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.437613 4972 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2eabfdfb-f2b5-458a-8f35-8c1aff548e91-util\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.750391 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerStarted","Data":"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9"} Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.755191 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" event={"ID":"2eabfdfb-f2b5-458a-8f35-8c1aff548e91","Type":"ContainerDied","Data":"aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56"} Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.755268 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa3296a9f69d6273640878dd1256d36f3581f9f9390d49da873dd1bbbd725d56" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.755286 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm" Feb 28 10:48:12 crc kubenswrapper[4972]: I0228 10:48:12.828535 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-txp88" podStartSLOduration=3.197191737 podStartE2EDuration="5.828508219s" podCreationTimestamp="2026-02-28 10:48:07 +0000 UTC" firstStartedPulling="2026-02-28 10:48:09.713860011 +0000 UTC m=+746.625867789" lastFinishedPulling="2026-02-28 10:48:12.345176493 +0000 UTC m=+749.257184271" observedRunningTime="2026-02-28 10:48:12.82317888 +0000 UTC m=+749.735186688" watchObservedRunningTime="2026-02-28 10:48:12.828508219 +0000 UTC m=+749.740515987" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.312737 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6"] Feb 28 10:48:16 crc kubenswrapper[4972]: E0228 10:48:16.313266 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="pull" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.313280 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="pull" Feb 28 10:48:16 crc kubenswrapper[4972]: E0228 10:48:16.313290 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="extract" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.313296 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="extract" Feb 28 10:48:16 crc kubenswrapper[4972]: E0228 10:48:16.313306 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="util" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.313311 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="util" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.313401 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eabfdfb-f2b5-458a-8f35-8c1aff548e91" containerName="extract" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.313870 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.315888 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.317951 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.318562 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vrgc9" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.335170 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6"] Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.502033 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fghrc\" (UniqueName: \"kubernetes.io/projected/d8bdef5d-8a08-4e53-bb43-2f721ed3570e-kube-api-access-fghrc\") pod \"nmstate-operator-75c5dccd6c-srsp6\" (UID: \"d8bdef5d-8a08-4e53-bb43-2f721ed3570e\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.603577 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fghrc\" (UniqueName: \"kubernetes.io/projected/d8bdef5d-8a08-4e53-bb43-2f721ed3570e-kube-api-access-fghrc\") pod \"nmstate-operator-75c5dccd6c-srsp6\" (UID: \"d8bdef5d-8a08-4e53-bb43-2f721ed3570e\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.626555 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fghrc\" (UniqueName: \"kubernetes.io/projected/d8bdef5d-8a08-4e53-bb43-2f721ed3570e-kube-api-access-fghrc\") pod \"nmstate-operator-75c5dccd6c-srsp6\" (UID: \"d8bdef5d-8a08-4e53-bb43-2f721ed3570e\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.633192 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" Feb 28 10:48:16 crc kubenswrapper[4972]: I0228 10:48:16.851936 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6"] Feb 28 10:48:17 crc kubenswrapper[4972]: I0228 10:48:17.822994 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" event={"ID":"d8bdef5d-8a08-4e53-bb43-2f721ed3570e","Type":"ContainerStarted","Data":"464ff2c7a252abe0d7dce7b7796700df1e7fd65f920c9daef00720386b6bad0e"} Feb 28 10:48:18 crc kubenswrapper[4972]: I0228 10:48:18.339199 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:18 crc kubenswrapper[4972]: I0228 10:48:18.339564 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:19 crc kubenswrapper[4972]: I0228 10:48:19.386900 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-txp88" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="registry-server" probeResult="failure" output=< Feb 28 10:48:19 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:48:19 crc kubenswrapper[4972]: > Feb 28 10:48:20 crc kubenswrapper[4972]: I0228 10:48:20.841325 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" event={"ID":"d8bdef5d-8a08-4e53-bb43-2f721ed3570e","Type":"ContainerStarted","Data":"acb2fd9cfd6114ad704046952fc24803dd0a05245f2d4603ee94d384d2824483"} Feb 28 10:48:20 crc kubenswrapper[4972]: I0228 10:48:20.860784 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-srsp6" podStartSLOduration=1.143758219 podStartE2EDuration="4.86076473s" podCreationTimestamp="2026-02-28 10:48:16 +0000 UTC" firstStartedPulling="2026-02-28 10:48:16.861513856 +0000 UTC m=+753.773521594" lastFinishedPulling="2026-02-28 10:48:20.578520357 +0000 UTC m=+757.490528105" observedRunningTime="2026-02-28 10:48:20.858830877 +0000 UTC m=+757.770838615" watchObservedRunningTime="2026-02-28 10:48:20.86076473 +0000 UTC m=+757.772772468" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.009296 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-sx977"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.011123 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.016788 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vwxr4" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.029241 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-2z62s"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.031099 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.033955 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-sx977"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.035847 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.075371 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-tt7f6"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.084225 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.109968 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-2z62s"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.152217 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqsrd\" (UniqueName: \"kubernetes.io/projected/a35c5570-d7ee-438e-8768-0daf418f5304-kube-api-access-rqsrd\") pod \"nmstate-metrics-69594cc75-sx977\" (UID: \"a35c5570-d7ee-438e-8768-0daf418f5304\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.152284 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbbcz\" (UniqueName: \"kubernetes.io/projected/b9a8aae0-4abb-4237-b2da-36542bac6c31-kube-api-access-qbbcz\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.152343 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9a8aae0-4abb-4237-b2da-36542bac6c31-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.173749 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.174480 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.180405 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.180738 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-c8xtk" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.181149 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.194033 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254232 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzws7\" (UniqueName: \"kubernetes.io/projected/d5746037-ce0c-44c8-9874-23ac38ad6509-kube-api-access-qzws7\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254536 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-dbus-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254730 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254805 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqsrd\" (UniqueName: \"kubernetes.io/projected/a35c5570-d7ee-438e-8768-0daf418f5304-kube-api-access-rqsrd\") pod \"nmstate-metrics-69594cc75-sx977\" (UID: \"a35c5570-d7ee-438e-8768-0daf418f5304\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254869 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-ovs-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254891 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbbcz\" (UniqueName: \"kubernetes.io/projected/b9a8aae0-4abb-4237-b2da-36542bac6c31-kube-api-access-qbbcz\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254937 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9a8aae0-4abb-4237-b2da-36542bac6c31-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.254966 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-nmstate-lock\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.255024 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pppg4\" (UniqueName: \"kubernetes.io/projected/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-kube-api-access-pppg4\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.255049 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5746037-ce0c-44c8-9874-23ac38ad6509-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.260807 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9a8aae0-4abb-4237-b2da-36542bac6c31-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.274113 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqsrd\" (UniqueName: \"kubernetes.io/projected/a35c5570-d7ee-438e-8768-0daf418f5304-kube-api-access-rqsrd\") pod \"nmstate-metrics-69594cc75-sx977\" (UID: \"a35c5570-d7ee-438e-8768-0daf418f5304\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.280256 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbbcz\" (UniqueName: \"kubernetes.io/projected/b9a8aae0-4abb-4237-b2da-36542bac6c31-kube-api-access-qbbcz\") pod \"nmstate-webhook-786f45cff4-2z62s\" (UID: \"b9a8aae0-4abb-4237-b2da-36542bac6c31\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356348 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-dbus-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356476 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356522 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-ovs-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356549 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-nmstate-lock\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356582 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pppg4\" (UniqueName: \"kubernetes.io/projected/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-kube-api-access-pppg4\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356614 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5746037-ce0c-44c8-9874-23ac38ad6509-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356660 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-ovs-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: E0228 10:48:26.356706 4972 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356721 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-nmstate-lock\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: E0228 10:48:26.356771 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert podName:d5746037-ce0c-44c8-9874-23ac38ad6509 nodeName:}" failed. No retries permitted until 2026-02-28 10:48:26.856749558 +0000 UTC m=+763.768757296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert") pod "nmstate-console-plugin-5dcbbd79cf-pdwzw" (UID: "d5746037-ce0c-44c8-9874-23ac38ad6509") : secret "plugin-serving-cert" not found Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356823 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzws7\" (UniqueName: \"kubernetes.io/projected/d5746037-ce0c-44c8-9874-23ac38ad6509-kube-api-access-qzws7\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.356823 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-dbus-socket\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.357701 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5746037-ce0c-44c8-9874-23ac38ad6509-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.367544 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-884c8bf7b-dtl6w"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.368357 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.371684 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.378369 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pppg4\" (UniqueName: \"kubernetes.io/projected/1013fa3e-26c1-4b74-a4f5-55c3b18fe34d-kube-api-access-pppg4\") pod \"nmstate-handler-tt7f6\" (UID: \"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d\") " pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.390314 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzws7\" (UniqueName: \"kubernetes.io/projected/d5746037-ce0c-44c8-9874-23ac38ad6509-kube-api-access-qzws7\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.391153 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-884c8bf7b-dtl6w"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.398892 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.410603 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:26 crc kubenswrapper[4972]: W0228 10:48:26.452323 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1013fa3e_26c1_4b74_a4f5_55c3b18fe34d.slice/crio-e5195ad9dbdc5e844603eba502191627412d063d518cb2708b89e3782fee4702 WatchSource:0}: Error finding container e5195ad9dbdc5e844603eba502191627412d063d518cb2708b89e3782fee4702: Status 404 returned error can't find the container with id e5195ad9dbdc5e844603eba502191627412d063d518cb2708b89e3782fee4702 Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464432 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-service-ca\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464516 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464560 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-trusted-ca-bundle\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464574 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-console-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464597 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-oauth-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464677 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-oauth-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.464742 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9tk7\" (UniqueName: \"kubernetes.io/projected/ff91b501-6437-460d-a952-aa5ee34d15b3-kube-api-access-n9tk7\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.566490 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9tk7\" (UniqueName: \"kubernetes.io/projected/ff91b501-6437-460d-a952-aa5ee34d15b3-kube-api-access-n9tk7\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.566895 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-service-ca\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.566937 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.566963 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-trusted-ca-bundle\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.567002 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-console-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.567024 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-oauth-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.567046 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-oauth-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.567945 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-service-ca\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.569869 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-trusted-ca-bundle\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.570218 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-console-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.570931 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ff91b501-6437-460d-a952-aa5ee34d15b3-oauth-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.573050 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-oauth-config\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.577363 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ff91b501-6437-460d-a952-aa5ee34d15b3-console-serving-cert\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.585025 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9tk7\" (UniqueName: \"kubernetes.io/projected/ff91b501-6437-460d-a952-aa5ee34d15b3-kube-api-access-n9tk7\") pod \"console-884c8bf7b-dtl6w\" (UID: \"ff91b501-6437-460d-a952-aa5ee34d15b3\") " pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.653288 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-2z62s"] Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.729225 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.808994 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-sx977"] Feb 28 10:48:26 crc kubenswrapper[4972]: W0228 10:48:26.818989 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda35c5570_d7ee_438e_8768_0daf418f5304.slice/crio-2e8ae812d37ee796cfb0b62cd0a5e19a2fdc0758031bfcd52a7c22d91dda2ad7 WatchSource:0}: Error finding container 2e8ae812d37ee796cfb0b62cd0a5e19a2fdc0758031bfcd52a7c22d91dda2ad7: Status 404 returned error can't find the container with id 2e8ae812d37ee796cfb0b62cd0a5e19a2fdc0758031bfcd52a7c22d91dda2ad7 Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.872984 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.882822 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5746037-ce0c-44c8-9874-23ac38ad6509-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-pdwzw\" (UID: \"d5746037-ce0c-44c8-9874-23ac38ad6509\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.896400 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" event={"ID":"a35c5570-d7ee-438e-8768-0daf418f5304","Type":"ContainerStarted","Data":"2e8ae812d37ee796cfb0b62cd0a5e19a2fdc0758031bfcd52a7c22d91dda2ad7"} Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.897579 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" event={"ID":"b9a8aae0-4abb-4237-b2da-36542bac6c31","Type":"ContainerStarted","Data":"6d5aef8c0c6fc33144a284d483c778cb530c81bdf269ea46e3f13a5318f8a980"} Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.898474 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tt7f6" event={"ID":"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d","Type":"ContainerStarted","Data":"e5195ad9dbdc5e844603eba502191627412d063d518cb2708b89e3782fee4702"} Feb 28 10:48:26 crc kubenswrapper[4972]: I0228 10:48:26.949639 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-884c8bf7b-dtl6w"] Feb 28 10:48:26 crc kubenswrapper[4972]: W0228 10:48:26.956232 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff91b501_6437_460d_a952_aa5ee34d15b3.slice/crio-9aeb944d1100e7ddb15e123bb40b509bfcd86b3c918641c9f50a8ee10ecf864f WatchSource:0}: Error finding container 9aeb944d1100e7ddb15e123bb40b509bfcd86b3c918641c9f50a8ee10ecf864f: Status 404 returned error can't find the container with id 9aeb944d1100e7ddb15e123bb40b509bfcd86b3c918641c9f50a8ee10ecf864f Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.096063 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.568145 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw"] Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.904918 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" event={"ID":"d5746037-ce0c-44c8-9874-23ac38ad6509","Type":"ContainerStarted","Data":"27ec0d8de8bd8a3d631663ec33ea6f5b1849cfb182e9f484b971681be416a823"} Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.906599 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-884c8bf7b-dtl6w" event={"ID":"ff91b501-6437-460d-a952-aa5ee34d15b3","Type":"ContainerStarted","Data":"ebb41430efe6393dd0fa8dc48cd3fc5581ae84f208698c7a8ce0634b3d95d22c"} Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.906631 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-884c8bf7b-dtl6w" event={"ID":"ff91b501-6437-460d-a952-aa5ee34d15b3","Type":"ContainerStarted","Data":"9aeb944d1100e7ddb15e123bb40b509bfcd86b3c918641c9f50a8ee10ecf864f"} Feb 28 10:48:27 crc kubenswrapper[4972]: I0228 10:48:27.934610 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-884c8bf7b-dtl6w" podStartSLOduration=1.93458592 podStartE2EDuration="1.93458592s" podCreationTimestamp="2026-02-28 10:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:48:27.925825576 +0000 UTC m=+764.837833324" watchObservedRunningTime="2026-02-28 10:48:27.93458592 +0000 UTC m=+764.846593658" Feb 28 10:48:28 crc kubenswrapper[4972]: I0228 10:48:28.382075 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:28 crc kubenswrapper[4972]: I0228 10:48:28.440372 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:28 crc kubenswrapper[4972]: I0228 10:48:28.621966 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:29 crc kubenswrapper[4972]: I0228 10:48:29.923644 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-txp88" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="registry-server" containerID="cri-o://c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9" gracePeriod=2 Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.312823 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.424235 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zhf7\" (UniqueName: \"kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7\") pod \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.424322 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content\") pod \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.424374 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities\") pod \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\" (UID: \"70f806dc-9d95-44c6-8a32-cecf2e921c3b\") " Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.425349 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities" (OuterVolumeSpecName: "utilities") pod "70f806dc-9d95-44c6-8a32-cecf2e921c3b" (UID: "70f806dc-9d95-44c6-8a32-cecf2e921c3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.432793 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7" (OuterVolumeSpecName: "kube-api-access-6zhf7") pod "70f806dc-9d95-44c6-8a32-cecf2e921c3b" (UID: "70f806dc-9d95-44c6-8a32-cecf2e921c3b"). InnerVolumeSpecName "kube-api-access-6zhf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.526037 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.526074 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zhf7\" (UniqueName: \"kubernetes.io/projected/70f806dc-9d95-44c6-8a32-cecf2e921c3b-kube-api-access-6zhf7\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.550958 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70f806dc-9d95-44c6-8a32-cecf2e921c3b" (UID: "70f806dc-9d95-44c6-8a32-cecf2e921c3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.627397 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f806dc-9d95-44c6-8a32-cecf2e921c3b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.936818 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" event={"ID":"b9a8aae0-4abb-4237-b2da-36542bac6c31","Type":"ContainerStarted","Data":"bf1be54f18dfb9c617fefc3eeb1c79e8997cafe742289516df8918dab50c7a4f"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.938430 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.941267 4972 generic.go:334] "Generic (PLEG): container finished" podID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerID="c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9" exitCode=0 Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.941349 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerDied","Data":"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.941405 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txp88" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.941435 4972 scope.go:117] "RemoveContainer" containerID="c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.941411 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txp88" event={"ID":"70f806dc-9d95-44c6-8a32-cecf2e921c3b","Type":"ContainerDied","Data":"f19b7f95b976af846e06725f815905ac283b89ceff7f0a30a0b482d879dc5941"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.944079 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tt7f6" event={"ID":"1013fa3e-26c1-4b74-a4f5-55c3b18fe34d","Type":"ContainerStarted","Data":"0e65d908768347751f10e9d0fb0242e5ee23d67ed6b7669c323441a47fb077e6"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.944277 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.946479 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" event={"ID":"a35c5570-d7ee-438e-8768-0daf418f5304","Type":"ContainerStarted","Data":"b1e4611c7f6fd0b28f2c00c8a8e086e25f104cc38bf4fe5b0b1fd2ac62fa87f1"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.948227 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" event={"ID":"d5746037-ce0c-44c8-9874-23ac38ad6509","Type":"ContainerStarted","Data":"378bf7f9bea1d211a3bc16a196f1d015af097e5a809a0205daf580cdbbaf07b9"} Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.960527 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" podStartSLOduration=1.449892407 podStartE2EDuration="4.960504267s" podCreationTimestamp="2026-02-28 10:48:26 +0000 UTC" firstStartedPulling="2026-02-28 10:48:26.661704606 +0000 UTC m=+763.573712344" lastFinishedPulling="2026-02-28 10:48:30.172316466 +0000 UTC m=+767.084324204" observedRunningTime="2026-02-28 10:48:30.957516783 +0000 UTC m=+767.869524541" watchObservedRunningTime="2026-02-28 10:48:30.960504267 +0000 UTC m=+767.872512005" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.967526 4972 scope.go:117] "RemoveContainer" containerID="de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.994727 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-tt7f6" podStartSLOduration=1.250769408 podStartE2EDuration="4.994704844s" podCreationTimestamp="2026-02-28 10:48:26 +0000 UTC" firstStartedPulling="2026-02-28 10:48:26.454735398 +0000 UTC m=+763.366743136" lastFinishedPulling="2026-02-28 10:48:30.198670834 +0000 UTC m=+767.110678572" observedRunningTime="2026-02-28 10:48:30.988021436 +0000 UTC m=+767.900029174" watchObservedRunningTime="2026-02-28 10:48:30.994704844 +0000 UTC m=+767.906712592" Feb 28 10:48:30 crc kubenswrapper[4972]: I0228 10:48:30.999042 4972 scope.go:117] "RemoveContainer" containerID="8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.012250 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-pdwzw" podStartSLOduration=2.422205046 podStartE2EDuration="5.012214753s" podCreationTimestamp="2026-02-28 10:48:26 +0000 UTC" firstStartedPulling="2026-02-28 10:48:27.584736127 +0000 UTC m=+764.496743905" lastFinishedPulling="2026-02-28 10:48:30.174745874 +0000 UTC m=+767.086753612" observedRunningTime="2026-02-28 10:48:31.004575079 +0000 UTC m=+767.916582837" watchObservedRunningTime="2026-02-28 10:48:31.012214753 +0000 UTC m=+767.924222501" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.029232 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.034230 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-txp88"] Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.035612 4972 scope.go:117] "RemoveContainer" containerID="c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9" Feb 28 10:48:31 crc kubenswrapper[4972]: E0228 10:48:31.036392 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9\": container with ID starting with c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9 not found: ID does not exist" containerID="c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.036431 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9"} err="failed to get container status \"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9\": rpc error: code = NotFound desc = could not find container \"c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9\": container with ID starting with c33227e9ce4845a10be85afd6296ca1a96310c004998b6b894873654c0388ed9 not found: ID does not exist" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.036484 4972 scope.go:117] "RemoveContainer" containerID="de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb" Feb 28 10:48:31 crc kubenswrapper[4972]: E0228 10:48:31.037170 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb\": container with ID starting with de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb not found: ID does not exist" containerID="de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.037214 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb"} err="failed to get container status \"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb\": rpc error: code = NotFound desc = could not find container \"de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb\": container with ID starting with de080a8c056afb56f5154dcf3c2fecf0d5dc00f6ec4a994cac20226a5624f2bb not found: ID does not exist" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.037243 4972 scope.go:117] "RemoveContainer" containerID="8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294" Feb 28 10:48:31 crc kubenswrapper[4972]: E0228 10:48:31.037499 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294\": container with ID starting with 8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294 not found: ID does not exist" containerID="8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.037527 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294"} err="failed to get container status \"8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294\": rpc error: code = NotFound desc = could not find container \"8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294\": container with ID starting with 8a7d4f74b45db78e61fe522c3c217fac016723907ecbf76bfd0dcbc2c70e6294 not found: ID does not exist" Feb 28 10:48:31 crc kubenswrapper[4972]: I0228 10:48:31.798398 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" path="/var/lib/kubelet/pods/70f806dc-9d95-44c6-8a32-cecf2e921c3b/volumes" Feb 28 10:48:32 crc kubenswrapper[4972]: I0228 10:48:32.970425 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" event={"ID":"a35c5570-d7ee-438e-8768-0daf418f5304","Type":"ContainerStarted","Data":"384ef222266e5834fd18cba18343d401aa2933887b6aa797c5e345304fef11bd"} Feb 28 10:48:32 crc kubenswrapper[4972]: I0228 10:48:32.997156 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-69594cc75-sx977" podStartSLOduration=2.324275413 podStartE2EDuration="7.997138229s" podCreationTimestamp="2026-02-28 10:48:25 +0000 UTC" firstStartedPulling="2026-02-28 10:48:26.82135391 +0000 UTC m=+763.733361658" lastFinishedPulling="2026-02-28 10:48:32.494216726 +0000 UTC m=+769.406224474" observedRunningTime="2026-02-28 10:48:32.996232534 +0000 UTC m=+769.908240292" watchObservedRunningTime="2026-02-28 10:48:32.997138229 +0000 UTC m=+769.909145967" Feb 28 10:48:36 crc kubenswrapper[4972]: I0228 10:48:36.443544 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-tt7f6" Feb 28 10:48:36 crc kubenswrapper[4972]: I0228 10:48:36.730328 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:36 crc kubenswrapper[4972]: I0228 10:48:36.730762 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:36 crc kubenswrapper[4972]: I0228 10:48:36.737079 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:37 crc kubenswrapper[4972]: I0228 10:48:37.004334 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-884c8bf7b-dtl6w" Feb 28 10:48:37 crc kubenswrapper[4972]: I0228 10:48:37.085377 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:48:46 crc kubenswrapper[4972]: I0228 10:48:46.411790 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-786f45cff4-2z62s" Feb 28 10:48:57 crc kubenswrapper[4972]: I0228 10:48:57.103353 4972 scope.go:117] "RemoveContainer" containerID="2b44db430ad9cb26d45f4ea4a21a594ac3d534096bd21c254cca436dbe975cdd" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.631356 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx"] Feb 28 10:49:00 crc kubenswrapper[4972]: E0228 10:49:00.632815 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="registry-server" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.632842 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="registry-server" Feb 28 10:49:00 crc kubenswrapper[4972]: E0228 10:49:00.632891 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="extract-content" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.632903 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="extract-content" Feb 28 10:49:00 crc kubenswrapper[4972]: E0228 10:49:00.632923 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="extract-utilities" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.632936 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="extract-utilities" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.633169 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f806dc-9d95-44c6-8a32-cecf2e921c3b" containerName="registry-server" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.634795 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.637397 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.640266 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx"] Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.738109 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.738177 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h89d\" (UniqueName: \"kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.738220 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.840501 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h89d\" (UniqueName: \"kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.840926 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.841143 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.841602 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.841865 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.866111 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h89d\" (UniqueName: \"kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:00 crc kubenswrapper[4972]: I0228 10:49:00.961584 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:01 crc kubenswrapper[4972]: I0228 10:49:01.215032 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx"] Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.148539 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-czzdx" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" containerID="cri-o://89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c" gracePeriod=15 Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.190739 4972 generic.go:334] "Generic (PLEG): container finished" podID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerID="4f4f27639f7c6bfe69a76e45f667bf6e757ea2c6c7f82656e49d193589b487f6" exitCode=0 Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.190800 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" event={"ID":"38feb8ea-eefb-4037-ba7a-b974e35ea6c3","Type":"ContainerDied","Data":"4f4f27639f7c6bfe69a76e45f667bf6e757ea2c6c7f82656e49d193589b487f6"} Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.190836 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" event={"ID":"38feb8ea-eefb-4037-ba7a-b974e35ea6c3","Type":"ContainerStarted","Data":"67279dab4939adfb6098ad92baf624b8eab1a382dc47d9893d9af8e6882ee6f4"} Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.192925 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.247314 4972 patch_prober.go:28] interesting pod/console-f9d7485db-czzdx container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.247387 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-czzdx" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.525289 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-czzdx_6505a931-94e8-45b7-8945-61f412bd5fd1/console/0.log" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.525826 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.670724 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.670895 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.670974 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.671013 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6bld\" (UniqueName: \"kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.671062 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.671243 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.671287 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config\") pod \"6505a931-94e8-45b7-8945-61f412bd5fd1\" (UID: \"6505a931-94e8-45b7-8945-61f412bd5fd1\") " Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.672797 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.672873 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.672960 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config" (OuterVolumeSpecName: "console-config") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.672975 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca" (OuterVolumeSpecName: "service-ca") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.680133 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld" (OuterVolumeSpecName: "kube-api-access-s6bld") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "kube-api-access-s6bld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.683804 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.684684 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6505a931-94e8-45b7-8945-61f412bd5fd1" (UID: "6505a931-94e8-45b7-8945-61f412bd5fd1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773562 4972 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773660 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6bld\" (UniqueName: \"kubernetes.io/projected/6505a931-94e8-45b7-8945-61f412bd5fd1-kube-api-access-s6bld\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773722 4972 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773742 4972 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773799 4972 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-console-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773824 4972 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6505a931-94e8-45b7-8945-61f412bd5fd1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:02 crc kubenswrapper[4972]: I0228 10:49:02.773842 4972 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6505a931-94e8-45b7-8945-61f412bd5fd1-service-ca\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205399 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-czzdx_6505a931-94e8-45b7-8945-61f412bd5fd1/console/0.log" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205497 4972 generic.go:334] "Generic (PLEG): container finished" podID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerID="89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c" exitCode=2 Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205551 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-czzdx" event={"ID":"6505a931-94e8-45b7-8945-61f412bd5fd1","Type":"ContainerDied","Data":"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c"} Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205585 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-czzdx" event={"ID":"6505a931-94e8-45b7-8945-61f412bd5fd1","Type":"ContainerDied","Data":"4cd378189333db82536e0be7c690cbcfbbdcca29d1b902bc87d9d5c0999e7dd6"} Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205606 4972 scope.go:117] "RemoveContainer" containerID="89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.205751 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-czzdx" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.247570 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.248367 4972 scope.go:117] "RemoveContainer" containerID="89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c" Feb 28 10:49:03 crc kubenswrapper[4972]: E0228 10:49:03.249210 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c\": container with ID starting with 89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c not found: ID does not exist" containerID="89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.249262 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c"} err="failed to get container status \"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c\": rpc error: code = NotFound desc = could not find container \"89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c\": container with ID starting with 89c163c1bb513b7808ad577437a2a182169761781f6bf2d3a003e9a2abcb046c not found: ID does not exist" Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.254690 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-czzdx"] Feb 28 10:49:03 crc kubenswrapper[4972]: I0228 10:49:03.800147 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" path="/var/lib/kubelet/pods/6505a931-94e8-45b7-8945-61f412bd5fd1/volumes" Feb 28 10:49:04 crc kubenswrapper[4972]: I0228 10:49:04.216975 4972 generic.go:334] "Generic (PLEG): container finished" podID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerID="205d9c1d29cbfedb4cdec3f62bcb74d84643bf895c30affe250c9da5c4c24c25" exitCode=0 Feb 28 10:49:04 crc kubenswrapper[4972]: I0228 10:49:04.217071 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" event={"ID":"38feb8ea-eefb-4037-ba7a-b974e35ea6c3","Type":"ContainerDied","Data":"205d9c1d29cbfedb4cdec3f62bcb74d84643bf895c30affe250c9da5c4c24c25"} Feb 28 10:49:05 crc kubenswrapper[4972]: I0228 10:49:05.228415 4972 generic.go:334] "Generic (PLEG): container finished" podID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerID="6478843b46a32980f1008f52bb5dc0b74c4a7e68497ce06d2cdfb2056b332592" exitCode=0 Feb 28 10:49:05 crc kubenswrapper[4972]: I0228 10:49:05.228499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" event={"ID":"38feb8ea-eefb-4037-ba7a-b974e35ea6c3","Type":"ContainerDied","Data":"6478843b46a32980f1008f52bb5dc0b74c4a7e68497ce06d2cdfb2056b332592"} Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.501526 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.637687 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle\") pod \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.637910 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h89d\" (UniqueName: \"kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d\") pod \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.637981 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util\") pod \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\" (UID: \"38feb8ea-eefb-4037-ba7a-b974e35ea6c3\") " Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.639030 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle" (OuterVolumeSpecName: "bundle") pod "38feb8ea-eefb-4037-ba7a-b974e35ea6c3" (UID: "38feb8ea-eefb-4037-ba7a-b974e35ea6c3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.645085 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d" (OuterVolumeSpecName: "kube-api-access-4h89d") pod "38feb8ea-eefb-4037-ba7a-b974e35ea6c3" (UID: "38feb8ea-eefb-4037-ba7a-b974e35ea6c3"). InnerVolumeSpecName "kube-api-access-4h89d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.655485 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util" (OuterVolumeSpecName: "util") pod "38feb8ea-eefb-4037-ba7a-b974e35ea6c3" (UID: "38feb8ea-eefb-4037-ba7a-b974e35ea6c3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.740157 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h89d\" (UniqueName: \"kubernetes.io/projected/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-kube-api-access-4h89d\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.740193 4972 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-util\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:06 crc kubenswrapper[4972]: I0228 10:49:06.740204 4972 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/38feb8ea-eefb-4037-ba7a-b974e35ea6c3-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:49:07 crc kubenswrapper[4972]: I0228 10:49:07.244216 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" event={"ID":"38feb8ea-eefb-4037-ba7a-b974e35ea6c3","Type":"ContainerDied","Data":"67279dab4939adfb6098ad92baf624b8eab1a382dc47d9893d9af8e6882ee6f4"} Feb 28 10:49:07 crc kubenswrapper[4972]: I0228 10:49:07.244286 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67279dab4939adfb6098ad92baf624b8eab1a382dc47d9893d9af8e6882ee6f4" Feb 28 10:49:07 crc kubenswrapper[4972]: I0228 10:49:07.244361 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.969485 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-57bd854678-2r552"] Feb 28 10:49:15 crc kubenswrapper[4972]: E0228 10:49:15.970354 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="util" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970366 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="util" Feb 28 10:49:15 crc kubenswrapper[4972]: E0228 10:49:15.970376 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="pull" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970382 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="pull" Feb 28 10:49:15 crc kubenswrapper[4972]: E0228 10:49:15.970395 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970401 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" Feb 28 10:49:15 crc kubenswrapper[4972]: E0228 10:49:15.970408 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="extract" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970414 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="extract" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970521 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="6505a931-94e8-45b7-8945-61f412bd5fd1" containerName="console" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970536 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="38feb8ea-eefb-4037-ba7a-b974e35ea6c3" containerName="extract" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.970924 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.976220 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.976489 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xslrd" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.977418 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.977522 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 28 10:49:15 crc kubenswrapper[4972]: I0228 10:49:15.978681 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.038000 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57bd854678-2r552"] Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.078301 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8924\" (UniqueName: \"kubernetes.io/projected/7d4fb02a-0341-49cf-947f-86fb93ac710b-kube-api-access-m8924\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.078358 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-apiservice-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.078448 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-webhook-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.179741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-webhook-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.179861 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8924\" (UniqueName: \"kubernetes.io/projected/7d4fb02a-0341-49cf-947f-86fb93ac710b-kube-api-access-m8924\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.179900 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-apiservice-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.191072 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-webhook-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.191072 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d4fb02a-0341-49cf-947f-86fb93ac710b-apiservice-cert\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.228974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8924\" (UniqueName: \"kubernetes.io/projected/7d4fb02a-0341-49cf-947f-86fb93ac710b-kube-api-access-m8924\") pod \"metallb-operator-controller-manager-57bd854678-2r552\" (UID: \"7d4fb02a-0341-49cf-947f-86fb93ac710b\") " pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.282537 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk"] Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.283751 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.292756 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.292967 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.293088 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-n55r7" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.293821 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.333572 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk"] Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.382776 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvntm\" (UniqueName: \"kubernetes.io/projected/ac415cf4-e2bc-471a-91aa-2752b88b2636-kube-api-access-cvntm\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.382830 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-webhook-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.382858 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-apiservice-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.483901 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvntm\" (UniqueName: \"kubernetes.io/projected/ac415cf4-e2bc-471a-91aa-2752b88b2636-kube-api-access-cvntm\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.483961 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-webhook-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.483994 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-apiservice-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.494363 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-webhook-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.494519 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac415cf4-e2bc-471a-91aa-2752b88b2636-apiservice-cert\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.500710 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvntm\" (UniqueName: \"kubernetes.io/projected/ac415cf4-e2bc-471a-91aa-2752b88b2636-kube-api-access-cvntm\") pod \"metallb-operator-webhook-server-5cdfc6748f-z4vkk\" (UID: \"ac415cf4-e2bc-471a-91aa-2752b88b2636\") " pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.606848 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57bd854678-2r552"] Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.625238 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:16 crc kubenswrapper[4972]: I0228 10:49:16.878645 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk"] Feb 28 10:49:17 crc kubenswrapper[4972]: I0228 10:49:17.350806 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" event={"ID":"7d4fb02a-0341-49cf-947f-86fb93ac710b","Type":"ContainerStarted","Data":"e471bd849c645c8fdaab0136c9430e61254e25786e39509799a75d9e5e1b4261"} Feb 28 10:49:17 crc kubenswrapper[4972]: I0228 10:49:17.352218 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" event={"ID":"ac415cf4-e2bc-471a-91aa-2752b88b2636","Type":"ContainerStarted","Data":"4ceebc377aca3751be2c4f3bbfed5ea804b37916a6e4382542a7d224530732f1"} Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.414077 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" event={"ID":"ac415cf4-e2bc-471a-91aa-2752b88b2636","Type":"ContainerStarted","Data":"e3c4fb482551bac26cda1a92afbf8b55b79e627b73f30ea0c6b293c18113c6a5"} Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.417292 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.418416 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" event={"ID":"7d4fb02a-0341-49cf-947f-86fb93ac710b","Type":"ContainerStarted","Data":"3f1987217f95f457ec55000fb06f8299670fd1a86a67b2a3450d0fcfb69b06ed"} Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.418794 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.444500 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" podStartSLOduration=1.892496258 podStartE2EDuration="6.444448076s" podCreationTimestamp="2026-02-28 10:49:16 +0000 UTC" firstStartedPulling="2026-02-28 10:49:16.894087904 +0000 UTC m=+813.806095642" lastFinishedPulling="2026-02-28 10:49:21.446039712 +0000 UTC m=+818.358047460" observedRunningTime="2026-02-28 10:49:22.442136264 +0000 UTC m=+819.354144002" watchObservedRunningTime="2026-02-28 10:49:22.444448076 +0000 UTC m=+819.356455814" Feb 28 10:49:22 crc kubenswrapper[4972]: I0228 10:49:22.465339 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" podStartSLOduration=2.725421439 podStartE2EDuration="7.465321137s" podCreationTimestamp="2026-02-28 10:49:15 +0000 UTC" firstStartedPulling="2026-02-28 10:49:16.627930505 +0000 UTC m=+813.539938243" lastFinishedPulling="2026-02-28 10:49:21.367830193 +0000 UTC m=+818.279837941" observedRunningTime="2026-02-28 10:49:22.463029096 +0000 UTC m=+819.375036834" watchObservedRunningTime="2026-02-28 10:49:22.465321137 +0000 UTC m=+819.377328875" Feb 28 10:49:36 crc kubenswrapper[4972]: I0228 10:49:36.632898 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5cdfc6748f-z4vkk" Feb 28 10:49:56 crc kubenswrapper[4972]: I0228 10:49:56.299560 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-57bd854678-2r552" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.057207 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.058624 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.065962 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.066104 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4d55k" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.069697 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-j7n9h"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.072635 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.076103 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.076426 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.085606 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.108375 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z24ml\" (UniqueName: \"kubernetes.io/projected/c4ddb606-f9be-489d-8c0b-5c89adbb265a-kube-api-access-z24ml\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.108941 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109284 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-conf\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109400 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109564 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109680 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-sockets\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109796 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-reloader\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.109897 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscgv\" (UniqueName: \"kubernetes.io/projected/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-kube-api-access-pscgv\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.110053 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-startup\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.187288 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-n2xm6"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.188905 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.191881 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.192088 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.192250 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.192303 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-fzmn5" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212585 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-startup\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212632 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212656 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z24ml\" (UniqueName: \"kubernetes.io/projected/c4ddb606-f9be-489d-8c0b-5c89adbb265a-kube-api-access-z24ml\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212678 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6qvv\" (UniqueName: \"kubernetes.io/projected/b387ac1b-669a-43f3-b336-7df69f886a1a-kube-api-access-m6qvv\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212696 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212719 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-conf\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212765 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212780 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b387ac1b-669a-43f3-b336-7df69f886a1a-metallb-excludel2\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212808 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212824 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-sockets\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212840 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-reloader\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.212856 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscgv\" (UniqueName: \"kubernetes.io/projected/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-kube-api-access-pscgv\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.214544 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-startup\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.214537 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-conf\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.214777 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.214862 4972 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.214911 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs podName:0f19a4a9-5554-4e38-a29b-1d1ffde9fc87 nodeName:}" failed. No retries permitted until 2026-02-28 10:49:57.71489572 +0000 UTC m=+854.626903468 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs") pod "frr-k8s-j7n9h" (UID: "0f19a4a9-5554-4e38-a29b-1d1ffde9fc87") : secret "frr-k8s-certs-secret" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.215058 4972 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.215148 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert podName:c4ddb606-f9be-489d-8c0b-5c89adbb265a nodeName:}" failed. No retries permitted until 2026-02-28 10:49:57.715120636 +0000 UTC m=+854.627128564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert") pod "frr-k8s-webhook-server-7f989f654f-c9c7m" (UID: "c4ddb606-f9be-489d-8c0b-5c89adbb265a") : secret "frr-k8s-webhook-server-cert" not found Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.220017 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-lxdv7"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.220712 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-frr-sockets\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.221200 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-reloader\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.223111 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.227539 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-lxdv7"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.234681 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.251182 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscgv\" (UniqueName: \"kubernetes.io/projected/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-kube-api-access-pscgv\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.262397 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z24ml\" (UniqueName: \"kubernetes.io/projected/c4ddb606-f9be-489d-8c0b-5c89adbb265a-kube-api-access-z24ml\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.314764 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-cert\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.314866 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.314930 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6qvv\" (UniqueName: \"kubernetes.io/projected/b387ac1b-669a-43f3-b336-7df69f886a1a-kube-api-access-m6qvv\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.315027 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.315065 4972 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.315169 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist podName:b387ac1b-669a-43f3-b336-7df69f886a1a nodeName:}" failed. No retries permitted until 2026-02-28 10:49:57.815142482 +0000 UTC m=+854.727150420 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist") pod "speaker-n2xm6" (UID: "b387ac1b-669a-43f3-b336-7df69f886a1a") : secret "metallb-memberlist" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.315185 4972 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.315241 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs podName:b387ac1b-669a-43f3-b336-7df69f886a1a nodeName:}" failed. No retries permitted until 2026-02-28 10:49:57.815222974 +0000 UTC m=+854.727230702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs") pod "speaker-n2xm6" (UID: "b387ac1b-669a-43f3-b336-7df69f886a1a") : secret "speaker-certs-secret" not found Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.315054 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7zt\" (UniqueName: \"kubernetes.io/projected/47d6aa51-c3aa-4f21-a02f-36663c003b7f-kube-api-access-vv7zt\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.315308 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b387ac1b-669a-43f3-b336-7df69f886a1a-metallb-excludel2\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.315332 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-metrics-certs\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.316142 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b387ac1b-669a-43f3-b336-7df69f886a1a-metallb-excludel2\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.333822 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6qvv\" (UniqueName: \"kubernetes.io/projected/b387ac1b-669a-43f3-b336-7df69f886a1a-kube-api-access-m6qvv\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.417083 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-cert\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.417212 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7zt\" (UniqueName: \"kubernetes.io/projected/47d6aa51-c3aa-4f21-a02f-36663c003b7f-kube-api-access-vv7zt\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.417256 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-metrics-certs\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.421798 4972 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.422217 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-metrics-certs\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.433110 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47d6aa51-c3aa-4f21-a02f-36663c003b7f-cert\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.440440 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7zt\" (UniqueName: \"kubernetes.io/projected/47d6aa51-c3aa-4f21-a02f-36663c003b7f-kube-api-access-vv7zt\") pod \"controller-86ddb6bd46-lxdv7\" (UID: \"47d6aa51-c3aa-4f21-a02f-36663c003b7f\") " pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.551703 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.722330 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.722835 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.727917 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f19a4a9-5554-4e38-a29b-1d1ffde9fc87-metrics-certs\") pod \"frr-k8s-j7n9h\" (UID: \"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87\") " pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.728081 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c4ddb606-f9be-489d-8c0b-5c89adbb265a-cert\") pod \"frr-k8s-webhook-server-7f989f654f-c9c7m\" (UID: \"c4ddb606-f9be-489d-8c0b-5c89adbb265a\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.734092 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.744659 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.824219 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.824343 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.824479 4972 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 28 10:49:57 crc kubenswrapper[4972]: E0228 10:49:57.824535 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist podName:b387ac1b-669a-43f3-b336-7df69f886a1a nodeName:}" failed. No retries permitted until 2026-02-28 10:49:58.824520513 +0000 UTC m=+855.736528251 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist") pod "speaker-n2xm6" (UID: "b387ac1b-669a-43f3-b336-7df69f886a1a") : secret "metallb-memberlist" not found Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.829692 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-metrics-certs\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.839045 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-lxdv7"] Feb 28 10:49:57 crc kubenswrapper[4972]: I0228 10:49:57.974914 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m"] Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.690504 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" event={"ID":"c4ddb606-f9be-489d-8c0b-5c89adbb265a","Type":"ContainerStarted","Data":"5a4a24c6c05a47cefe5c789889fa02d90745714e2f8000b9af3b7713034103df"} Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.693517 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-lxdv7" event={"ID":"47d6aa51-c3aa-4f21-a02f-36663c003b7f","Type":"ContainerStarted","Data":"79533d447b2cd3ee9443225b2ac950ddcb9966d5e35e14cd49f1f49b7f93f027"} Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.693597 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-lxdv7" event={"ID":"47d6aa51-c3aa-4f21-a02f-36663c003b7f","Type":"ContainerStarted","Data":"326239d6fb9253bc6e0ed65fbf348d97da881efe711863437b51ef066fff9c8f"} Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.693628 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.693649 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-lxdv7" event={"ID":"47d6aa51-c3aa-4f21-a02f-36663c003b7f","Type":"ContainerStarted","Data":"f2f3bd90ec6f77bba24ddb092ff241974a93cecf7296f02ba62a0505ca8321f7"} Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.694723 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"77c5917f050deed0642a1001a437db868cc06439d9fa6ede7996f54cb45c66e7"} Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.720585 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-lxdv7" podStartSLOduration=1.7205527790000001 podStartE2EDuration="1.720552779s" podCreationTimestamp="2026-02-28 10:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:49:58.719576933 +0000 UTC m=+855.631584691" watchObservedRunningTime="2026-02-28 10:49:58.720552779 +0000 UTC m=+855.632560517" Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.838482 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:58 crc kubenswrapper[4972]: I0228 10:49:58.846686 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b387ac1b-669a-43f3-b336-7df69f886a1a-memberlist\") pod \"speaker-n2xm6\" (UID: \"b387ac1b-669a-43f3-b336-7df69f886a1a\") " pod="metallb-system/speaker-n2xm6" Feb 28 10:49:59 crc kubenswrapper[4972]: I0228 10:49:59.035089 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-n2xm6" Feb 28 10:49:59 crc kubenswrapper[4972]: I0228 10:49:59.714154 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n2xm6" event={"ID":"b387ac1b-669a-43f3-b336-7df69f886a1a","Type":"ContainerStarted","Data":"81647bb997d0408d6922d0b532217d3d2bb35c5dcc092f1b043f6c415eed6f32"} Feb 28 10:49:59 crc kubenswrapper[4972]: I0228 10:49:59.714680 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n2xm6" event={"ID":"b387ac1b-669a-43f3-b336-7df69f886a1a","Type":"ContainerStarted","Data":"697c573867bcf7dfbd856b3e9b02579fd0c375679dbc46aee7f5b3bf7d4e925b"} Feb 28 10:49:59 crc kubenswrapper[4972]: I0228 10:49:59.714719 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-n2xm6" event={"ID":"b387ac1b-669a-43f3-b336-7df69f886a1a","Type":"ContainerStarted","Data":"0832cca3fbbdc1c42276bb7a20a9aa57003cf30cc838d513c9d81140a3e63b15"} Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.137974 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537930-8wvgq"] Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.139439 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.145203 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.145521 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.145815 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.151813 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537930-8wvgq"] Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.288449 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49c6c\" (UniqueName: \"kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c\") pod \"auto-csr-approver-29537930-8wvgq\" (UID: \"35a163a1-a70c-4052-bc4c-d54cf60b4613\") " pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.389884 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49c6c\" (UniqueName: \"kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c\") pod \"auto-csr-approver-29537930-8wvgq\" (UID: \"35a163a1-a70c-4052-bc4c-d54cf60b4613\") " pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.411962 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49c6c\" (UniqueName: \"kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c\") pod \"auto-csr-approver-29537930-8wvgq\" (UID: \"35a163a1-a70c-4052-bc4c-d54cf60b4613\") " pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.500044 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.722875 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-n2xm6" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.747900 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-n2xm6" podStartSLOduration=3.747880912 podStartE2EDuration="3.747880912s" podCreationTimestamp="2026-02-28 10:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:50:00.745943391 +0000 UTC m=+857.657951149" watchObservedRunningTime="2026-02-28 10:50:00.747880912 +0000 UTC m=+857.659888650" Feb 28 10:50:00 crc kubenswrapper[4972]: I0228 10:50:00.839932 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537930-8wvgq"] Feb 28 10:50:01 crc kubenswrapper[4972]: I0228 10:50:01.731772 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" event={"ID":"35a163a1-a70c-4052-bc4c-d54cf60b4613","Type":"ContainerStarted","Data":"b3d9eb96d900d1719dce5dd2c07013ff3d1eff6198f891458ee9d6e867b1f510"} Feb 28 10:50:02 crc kubenswrapper[4972]: I0228 10:50:02.741934 4972 generic.go:334] "Generic (PLEG): container finished" podID="35a163a1-a70c-4052-bc4c-d54cf60b4613" containerID="bbc503054e17d6b71e191c0971b7edc76ab645da74b4e9634eef321953ea4704" exitCode=0 Feb 28 10:50:02 crc kubenswrapper[4972]: I0228 10:50:02.742019 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" event={"ID":"35a163a1-a70c-4052-bc4c-d54cf60b4613","Type":"ContainerDied","Data":"bbc503054e17d6b71e191c0971b7edc76ab645da74b4e9634eef321953ea4704"} Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.252082 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.362144 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49c6c\" (UniqueName: \"kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c\") pod \"35a163a1-a70c-4052-bc4c-d54cf60b4613\" (UID: \"35a163a1-a70c-4052-bc4c-d54cf60b4613\") " Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.376778 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c" (OuterVolumeSpecName: "kube-api-access-49c6c") pod "35a163a1-a70c-4052-bc4c-d54cf60b4613" (UID: "35a163a1-a70c-4052-bc4c-d54cf60b4613"). InnerVolumeSpecName "kube-api-access-49c6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.464336 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49c6c\" (UniqueName: \"kubernetes.io/projected/35a163a1-a70c-4052-bc4c-d54cf60b4613-kube-api-access-49c6c\") on node \"crc\" DevicePath \"\"" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.776335 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" event={"ID":"35a163a1-a70c-4052-bc4c-d54cf60b4613","Type":"ContainerDied","Data":"b3d9eb96d900d1719dce5dd2c07013ff3d1eff6198f891458ee9d6e867b1f510"} Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.776827 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3d9eb96d900d1719dce5dd2c07013ff3d1eff6198f891458ee9d6e867b1f510" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.776396 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537930-8wvgq" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.779053 4972 generic.go:334] "Generic (PLEG): container finished" podID="0f19a4a9-5554-4e38-a29b-1d1ffde9fc87" containerID="4a5daa6bf8cd96cd0cd2eca6b1c3f4ff4fb4c0f812a9ffeba4cb7d7ed3a56f4a" exitCode=0 Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.779165 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerDied","Data":"4a5daa6bf8cd96cd0cd2eca6b1c3f4ff4fb4c0f812a9ffeba4cb7d7ed3a56f4a"} Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.782157 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" event={"ID":"c4ddb606-f9be-489d-8c0b-5c89adbb265a","Type":"ContainerStarted","Data":"5bb657b6bd6bd45ea91e81e2683ffed7a10b2b6d42fd1e9e0f287de6e99d4a27"} Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.782421 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:50:05 crc kubenswrapper[4972]: I0228 10:50:05.852845 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" podStartSLOduration=1.545660273 podStartE2EDuration="8.852821284s" podCreationTimestamp="2026-02-28 10:49:57 +0000 UTC" firstStartedPulling="2026-02-28 10:49:57.993811599 +0000 UTC m=+854.905819337" lastFinishedPulling="2026-02-28 10:50:05.30097261 +0000 UTC m=+862.212980348" observedRunningTime="2026-02-28 10:50:05.847907154 +0000 UTC m=+862.759914892" watchObservedRunningTime="2026-02-28 10:50:05.852821284 +0000 UTC m=+862.764829022" Feb 28 10:50:06 crc kubenswrapper[4972]: I0228 10:50:06.312313 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537924-h2gm7"] Feb 28 10:50:06 crc kubenswrapper[4972]: I0228 10:50:06.315561 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537924-h2gm7"] Feb 28 10:50:06 crc kubenswrapper[4972]: I0228 10:50:06.794053 4972 generic.go:334] "Generic (PLEG): container finished" podID="0f19a4a9-5554-4e38-a29b-1d1ffde9fc87" containerID="5999a38b5eb637332d39f2a885454b38a2f8a037046d68d6986c8a1b1b16f0a9" exitCode=0 Feb 28 10:50:06 crc kubenswrapper[4972]: I0228 10:50:06.794120 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerDied","Data":"5999a38b5eb637332d39f2a885454b38a2f8a037046d68d6986c8a1b1b16f0a9"} Feb 28 10:50:07 crc kubenswrapper[4972]: I0228 10:50:07.798492 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fcdc4df-a299-4e3c-970b-47c4e01aa51e" path="/var/lib/kubelet/pods/1fcdc4df-a299-4e3c-970b-47c4e01aa51e/volumes" Feb 28 10:50:07 crc kubenswrapper[4972]: I0228 10:50:07.802829 4972 generic.go:334] "Generic (PLEG): container finished" podID="0f19a4a9-5554-4e38-a29b-1d1ffde9fc87" containerID="63ad2df45e57f2757ba22f00777a2f87c05f73a40cb39b35c6254348dbfe8888" exitCode=0 Feb 28 10:50:07 crc kubenswrapper[4972]: I0228 10:50:07.802885 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerDied","Data":"63ad2df45e57f2757ba22f00777a2f87c05f73a40cb39b35c6254348dbfe8888"} Feb 28 10:50:08 crc kubenswrapper[4972]: I0228 10:50:08.825076 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"cf4a2973749b41c13b6cc7cddad0642e4cb1f86049ec58fe382aae76200d341b"} Feb 28 10:50:08 crc kubenswrapper[4972]: I0228 10:50:08.825948 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"1ffb2faf30dda8fd21202ce86fd6a60564e0ef8e4e673cd59de0cf32aeaca3bf"} Feb 28 10:50:08 crc kubenswrapper[4972]: I0228 10:50:08.825965 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"5fd6e8f3eef7a97661e4f63307946fb8ed862c65782154abb302380b584fdf75"} Feb 28 10:50:08 crc kubenswrapper[4972]: I0228 10:50:08.825974 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"95b51d4edefd635944d7f02ed56c767f02737ad16c73b843ab4c72d69933f974"} Feb 28 10:50:08 crc kubenswrapper[4972]: I0228 10:50:08.825983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"3050ef502bb8582e0e2c84894045bb4b0d2d5a97e9c5c9c042e18dba97bbe0b1"} Feb 28 10:50:09 crc kubenswrapper[4972]: I0228 10:50:09.039104 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-n2xm6" Feb 28 10:50:09 crc kubenswrapper[4972]: I0228 10:50:09.842374 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7n9h" event={"ID":"0f19a4a9-5554-4e38-a29b-1d1ffde9fc87","Type":"ContainerStarted","Data":"4ee3f4f0ec3c7a940a26c3b4a0afd9dfb277760aafa67015508646b098b96a12"} Feb 28 10:50:09 crc kubenswrapper[4972]: I0228 10:50:09.842931 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:50:09 crc kubenswrapper[4972]: I0228 10:50:09.905691 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-j7n9h" podStartSLOduration=5.515316231 podStartE2EDuration="12.905657171s" podCreationTimestamp="2026-02-28 10:49:57 +0000 UTC" firstStartedPulling="2026-02-28 10:49:57.902008482 +0000 UTC m=+854.814016220" lastFinishedPulling="2026-02-28 10:50:05.292349422 +0000 UTC m=+862.204357160" observedRunningTime="2026-02-28 10:50:09.878489643 +0000 UTC m=+866.790497391" watchObservedRunningTime="2026-02-28 10:50:09.905657171 +0000 UTC m=+866.817664919" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.914222 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:11 crc kubenswrapper[4972]: E0228 10:50:11.914780 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a163a1-a70c-4052-bc4c-d54cf60b4613" containerName="oc" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.914806 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a163a1-a70c-4052-bc4c-d54cf60b4613" containerName="oc" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.915102 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a163a1-a70c-4052-bc4c-d54cf60b4613" containerName="oc" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.915929 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.918711 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.919840 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.921735 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xgr6t" Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.943854 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:11 crc kubenswrapper[4972]: I0228 10:50:11.971224 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkkjp\" (UniqueName: \"kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp\") pod \"openstack-operator-index-844px\" (UID: \"0968221f-c068-4bf3-9436-9547a37f3a43\") " pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.072838 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkkjp\" (UniqueName: \"kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp\") pod \"openstack-operator-index-844px\" (UID: \"0968221f-c068-4bf3-9436-9547a37f3a43\") " pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.098521 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkkjp\" (UniqueName: \"kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp\") pod \"openstack-operator-index-844px\" (UID: \"0968221f-c068-4bf3-9436-9547a37f3a43\") " pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.238016 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.616971 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:12 crc kubenswrapper[4972]: W0228 10:50:12.622210 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0968221f_c068_4bf3_9436_9547a37f3a43.slice/crio-3a298c3d3c806b60cb3e8c3d2371ae54da55a11a0234fc455d79df2ec5e46f80 WatchSource:0}: Error finding container 3a298c3d3c806b60cb3e8c3d2371ae54da55a11a0234fc455d79df2ec5e46f80: Status 404 returned error can't find the container with id 3a298c3d3c806b60cb3e8c3d2371ae54da55a11a0234fc455d79df2ec5e46f80 Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.745590 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.789031 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:50:12 crc kubenswrapper[4972]: I0228 10:50:12.867071 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-844px" event={"ID":"0968221f-c068-4bf3-9436-9547a37f3a43","Type":"ContainerStarted","Data":"3a298c3d3c806b60cb3e8c3d2371ae54da55a11a0234fc455d79df2ec5e46f80"} Feb 28 10:50:15 crc kubenswrapper[4972]: I0228 10:50:15.281436 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:15 crc kubenswrapper[4972]: I0228 10:50:15.884579 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pvw8x"] Feb 28 10:50:15 crc kubenswrapper[4972]: I0228 10:50:15.885639 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:15 crc kubenswrapper[4972]: I0228 10:50:15.900817 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pvw8x"] Feb 28 10:50:15 crc kubenswrapper[4972]: I0228 10:50:15.933997 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dcfp\" (UniqueName: \"kubernetes.io/projected/8e82abb6-3796-42c1-bfe7-688f9117943f-kube-api-access-7dcfp\") pod \"openstack-operator-index-pvw8x\" (UID: \"8e82abb6-3796-42c1-bfe7-688f9117943f\") " pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.034994 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dcfp\" (UniqueName: \"kubernetes.io/projected/8e82abb6-3796-42c1-bfe7-688f9117943f-kube-api-access-7dcfp\") pod \"openstack-operator-index-pvw8x\" (UID: \"8e82abb6-3796-42c1-bfe7-688f9117943f\") " pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.066704 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dcfp\" (UniqueName: \"kubernetes.io/projected/8e82abb6-3796-42c1-bfe7-688f9117943f-kube-api-access-7dcfp\") pod \"openstack-operator-index-pvw8x\" (UID: \"8e82abb6-3796-42c1-bfe7-688f9117943f\") " pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.210058 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.450118 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pvw8x"] Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.920004 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-844px" event={"ID":"0968221f-c068-4bf3-9436-9547a37f3a43","Type":"ContainerStarted","Data":"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6"} Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.920174 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-844px" podUID="0968221f-c068-4bf3-9436-9547a37f3a43" containerName="registry-server" containerID="cri-o://494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6" gracePeriod=2 Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.922263 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pvw8x" event={"ID":"8e82abb6-3796-42c1-bfe7-688f9117943f","Type":"ContainerStarted","Data":"c0a280fe0e98e6934b09771a9eef3f1e120620713791f520b79fddbcb167c182"} Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.922294 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pvw8x" event={"ID":"8e82abb6-3796-42c1-bfe7-688f9117943f","Type":"ContainerStarted","Data":"bbcc834fbb997d0f385ced75081a3d0a7a065b3f4cbe984d3236abe0193f56a5"} Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.947378 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-844px" podStartSLOduration=2.6716993540000002 podStartE2EDuration="5.947360661s" podCreationTimestamp="2026-02-28 10:50:11 +0000 UTC" firstStartedPulling="2026-02-28 10:50:12.625374184 +0000 UTC m=+869.537381922" lastFinishedPulling="2026-02-28 10:50:15.901035491 +0000 UTC m=+872.813043229" observedRunningTime="2026-02-28 10:50:16.944041723 +0000 UTC m=+873.856049501" watchObservedRunningTime="2026-02-28 10:50:16.947360661 +0000 UTC m=+873.859368399" Feb 28 10:50:16 crc kubenswrapper[4972]: I0228 10:50:16.970366 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pvw8x" podStartSLOduration=1.883065201 podStartE2EDuration="1.970343618s" podCreationTimestamp="2026-02-28 10:50:15 +0000 UTC" firstStartedPulling="2026-02-28 10:50:16.455591797 +0000 UTC m=+873.367599525" lastFinishedPulling="2026-02-28 10:50:16.542870194 +0000 UTC m=+873.454877942" observedRunningTime="2026-02-28 10:50:16.965211173 +0000 UTC m=+873.877218921" watchObservedRunningTime="2026-02-28 10:50:16.970343618 +0000 UTC m=+873.882351376" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.300194 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.358272 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkkjp\" (UniqueName: \"kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp\") pod \"0968221f-c068-4bf3-9436-9547a37f3a43\" (UID: \"0968221f-c068-4bf3-9436-9547a37f3a43\") " Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.366664 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp" (OuterVolumeSpecName: "kube-api-access-nkkjp") pod "0968221f-c068-4bf3-9436-9547a37f3a43" (UID: "0968221f-c068-4bf3-9436-9547a37f3a43"). InnerVolumeSpecName "kube-api-access-nkkjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.460479 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkkjp\" (UniqueName: \"kubernetes.io/projected/0968221f-c068-4bf3-9436-9547a37f3a43-kube-api-access-nkkjp\") on node \"crc\" DevicePath \"\"" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.556559 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-lxdv7" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.740303 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-c9c7m" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.749664 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-j7n9h" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.930268 4972 generic.go:334] "Generic (PLEG): container finished" podID="0968221f-c068-4bf3-9436-9547a37f3a43" containerID="494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6" exitCode=0 Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.930680 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-844px" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.931104 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-844px" event={"ID":"0968221f-c068-4bf3-9436-9547a37f3a43","Type":"ContainerDied","Data":"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6"} Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.931141 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-844px" event={"ID":"0968221f-c068-4bf3-9436-9547a37f3a43","Type":"ContainerDied","Data":"3a298c3d3c806b60cb3e8c3d2371ae54da55a11a0234fc455d79df2ec5e46f80"} Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.931162 4972 scope.go:117] "RemoveContainer" containerID="494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.949004 4972 scope.go:117] "RemoveContainer" containerID="494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6" Feb 28 10:50:17 crc kubenswrapper[4972]: E0228 10:50:17.949627 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6\": container with ID starting with 494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6 not found: ID does not exist" containerID="494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.949681 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6"} err="failed to get container status \"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6\": rpc error: code = NotFound desc = could not find container \"494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6\": container with ID starting with 494cb42929063f32a76a2034e99a9a5b1c96f274a18134e84315b0304e91f5e6 not found: ID does not exist" Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.961525 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:17 crc kubenswrapper[4972]: I0228 10:50:17.977820 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-844px"] Feb 28 10:50:19 crc kubenswrapper[4972]: I0228 10:50:19.798734 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0968221f-c068-4bf3-9436-9547a37f3a43" path="/var/lib/kubelet/pods/0968221f-c068-4bf3-9436-9547a37f3a43/volumes" Feb 28 10:50:26 crc kubenswrapper[4972]: I0228 10:50:26.211337 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:26 crc kubenswrapper[4972]: I0228 10:50:26.214577 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:26 crc kubenswrapper[4972]: I0228 10:50:26.250320 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.038414 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-pvw8x" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.726218 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph"] Feb 28 10:50:27 crc kubenswrapper[4972]: E0228 10:50:27.728095 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0968221f-c068-4bf3-9436-9547a37f3a43" containerName="registry-server" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.728207 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0968221f-c068-4bf3-9436-9547a37f3a43" containerName="registry-server" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.728433 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0968221f-c068-4bf3-9436-9547a37f3a43" containerName="registry-server" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.729645 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.733697 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zvhrm" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.753831 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph"] Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.838919 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.839365 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnn8b\" (UniqueName: \"kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.839567 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.941799 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.942331 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.942548 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.942852 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.942859 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnn8b\" (UniqueName: \"kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:27 crc kubenswrapper[4972]: I0228 10:50:27.975792 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnn8b\" (UniqueName: \"kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b\") pod \"38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:28 crc kubenswrapper[4972]: I0228 10:50:28.057158 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:28 crc kubenswrapper[4972]: I0228 10:50:28.475909 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph"] Feb 28 10:50:28 crc kubenswrapper[4972]: W0228 10:50:28.482940 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c8c46d3_deb0_4552_9e73_d565dd09da2a.slice/crio-0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b WatchSource:0}: Error finding container 0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b: Status 404 returned error can't find the container with id 0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b Feb 28 10:50:28 crc kubenswrapper[4972]: I0228 10:50:28.890328 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:50:28 crc kubenswrapper[4972]: I0228 10:50:28.890391 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:50:29 crc kubenswrapper[4972]: I0228 10:50:29.021672 4972 generic.go:334] "Generic (PLEG): container finished" podID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerID="b2dffe337b99640d2df11eaaf3bf10d6ffa8911aa52c52203d04ebee5eba808d" exitCode=0 Feb 28 10:50:29 crc kubenswrapper[4972]: I0228 10:50:29.021725 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" event={"ID":"4c8c46d3-deb0-4552-9e73-d565dd09da2a","Type":"ContainerDied","Data":"b2dffe337b99640d2df11eaaf3bf10d6ffa8911aa52c52203d04ebee5eba808d"} Feb 28 10:50:29 crc kubenswrapper[4972]: I0228 10:50:29.021753 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" event={"ID":"4c8c46d3-deb0-4552-9e73-d565dd09da2a","Type":"ContainerStarted","Data":"0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b"} Feb 28 10:50:30 crc kubenswrapper[4972]: I0228 10:50:30.029886 4972 generic.go:334] "Generic (PLEG): container finished" podID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerID="a81c1a81677518218fb09b9d4072e960e504783850a125364f754d88c43db607" exitCode=0 Feb 28 10:50:30 crc kubenswrapper[4972]: I0228 10:50:30.029984 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" event={"ID":"4c8c46d3-deb0-4552-9e73-d565dd09da2a","Type":"ContainerDied","Data":"a81c1a81677518218fb09b9d4072e960e504783850a125364f754d88c43db607"} Feb 28 10:50:31 crc kubenswrapper[4972]: I0228 10:50:31.042560 4972 generic.go:334] "Generic (PLEG): container finished" podID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerID="acf3369776c54ac63a2f8b1c0907a5aa5d6f77bb1db8955ab8f8898979df0c06" exitCode=0 Feb 28 10:50:31 crc kubenswrapper[4972]: I0228 10:50:31.042616 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" event={"ID":"4c8c46d3-deb0-4552-9e73-d565dd09da2a","Type":"ContainerDied","Data":"acf3369776c54ac63a2f8b1c0907a5aa5d6f77bb1db8955ab8f8898979df0c06"} Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.295084 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.410053 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util\") pod \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.410169 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnn8b\" (UniqueName: \"kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b\") pod \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.410195 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle\") pod \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\" (UID: \"4c8c46d3-deb0-4552-9e73-d565dd09da2a\") " Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.411086 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle" (OuterVolumeSpecName: "bundle") pod "4c8c46d3-deb0-4552-9e73-d565dd09da2a" (UID: "4c8c46d3-deb0-4552-9e73-d565dd09da2a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.418142 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b" (OuterVolumeSpecName: "kube-api-access-pnn8b") pod "4c8c46d3-deb0-4552-9e73-d565dd09da2a" (UID: "4c8c46d3-deb0-4552-9e73-d565dd09da2a"). InnerVolumeSpecName "kube-api-access-pnn8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.437481 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util" (OuterVolumeSpecName: "util") pod "4c8c46d3-deb0-4552-9e73-d565dd09da2a" (UID: "4c8c46d3-deb0-4552-9e73-d565dd09da2a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.512735 4972 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-util\") on node \"crc\" DevicePath \"\"" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.512840 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnn8b\" (UniqueName: \"kubernetes.io/projected/4c8c46d3-deb0-4552-9e73-d565dd09da2a-kube-api-access-pnn8b\") on node \"crc\" DevicePath \"\"" Feb 28 10:50:32 crc kubenswrapper[4972]: I0228 10:50:32.512866 4972 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c8c46d3-deb0-4552-9e73-d565dd09da2a-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:50:33 crc kubenswrapper[4972]: I0228 10:50:33.066595 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" event={"ID":"4c8c46d3-deb0-4552-9e73-d565dd09da2a","Type":"ContainerDied","Data":"0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b"} Feb 28 10:50:33 crc kubenswrapper[4972]: I0228 10:50:33.066692 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c251c4f0bffe738f6a82cda070df726db7a0aef735cd213451357aec2f16f9b" Feb 28 10:50:33 crc kubenswrapper[4972]: I0228 10:50:33.066741 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.121182 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x"] Feb 28 10:50:40 crc kubenswrapper[4972]: E0228 10:50:40.122070 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="extract" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.122082 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="extract" Feb 28 10:50:40 crc kubenswrapper[4972]: E0228 10:50:40.122100 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="util" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.122106 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="util" Feb 28 10:50:40 crc kubenswrapper[4972]: E0228 10:50:40.122116 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="pull" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.122122 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="pull" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.122266 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8c46d3-deb0-4552-9e73-d565dd09da2a" containerName="extract" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.122763 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.134857 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-kcx4l" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.163283 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x"] Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.223880 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgdmz\" (UniqueName: \"kubernetes.io/projected/cf37191c-db05-439a-9353-1e15f8147289-kube-api-access-pgdmz\") pod \"openstack-operator-controller-init-6cdcc94fb7-qfx4x\" (UID: \"cf37191c-db05-439a-9353-1e15f8147289\") " pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.325224 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgdmz\" (UniqueName: \"kubernetes.io/projected/cf37191c-db05-439a-9353-1e15f8147289-kube-api-access-pgdmz\") pod \"openstack-operator-controller-init-6cdcc94fb7-qfx4x\" (UID: \"cf37191c-db05-439a-9353-1e15f8147289\") " pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.352074 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgdmz\" (UniqueName: \"kubernetes.io/projected/cf37191c-db05-439a-9353-1e15f8147289-kube-api-access-pgdmz\") pod \"openstack-operator-controller-init-6cdcc94fb7-qfx4x\" (UID: \"cf37191c-db05-439a-9353-1e15f8147289\") " pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.445725 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:40 crc kubenswrapper[4972]: I0228 10:50:40.712974 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x"] Feb 28 10:50:41 crc kubenswrapper[4972]: I0228 10:50:41.136279 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" event={"ID":"cf37191c-db05-439a-9353-1e15f8147289","Type":"ContainerStarted","Data":"da4ed99a2ddf185bca1304d2aea53005ff81ab433de244a189d52a1743349a87"} Feb 28 10:50:46 crc kubenswrapper[4972]: I0228 10:50:46.205066 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" event={"ID":"cf37191c-db05-439a-9353-1e15f8147289","Type":"ContainerStarted","Data":"807e677b4c06a1cef7d20625e92f7b418b76ac4acfc104922043765b0e56ce53"} Feb 28 10:50:46 crc kubenswrapper[4972]: I0228 10:50:46.205961 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:46 crc kubenswrapper[4972]: I0228 10:50:46.266849 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" podStartSLOduration=1.869348163 podStartE2EDuration="6.266817265s" podCreationTimestamp="2026-02-28 10:50:40 +0000 UTC" firstStartedPulling="2026-02-28 10:50:40.734385814 +0000 UTC m=+897.646393552" lastFinishedPulling="2026-02-28 10:50:45.131854896 +0000 UTC m=+902.043862654" observedRunningTime="2026-02-28 10:50:46.25695084 +0000 UTC m=+903.168958608" watchObservedRunningTime="2026-02-28 10:50:46.266817265 +0000 UTC m=+903.178825043" Feb 28 10:50:50 crc kubenswrapper[4972]: I0228 10:50:50.449622 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6cdcc94fb7-qfx4x" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.249730 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.251326 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.280969 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.386864 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.386968 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxqzz\" (UniqueName: \"kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.386994 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.488159 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.488260 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxqzz\" (UniqueName: \"kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.488283 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.488713 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.488761 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.510155 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxqzz\" (UniqueName: \"kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz\") pod \"certified-operators-f6vm2\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:56 crc kubenswrapper[4972]: I0228 10:50:56.576197 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:50:57 crc kubenswrapper[4972]: I0228 10:50:57.104711 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:50:57 crc kubenswrapper[4972]: I0228 10:50:57.208057 4972 scope.go:117] "RemoveContainer" containerID="9fe3be65c4e68afcb7b9dfa412fad3a1caf9de652a625e28d6a44609f7d5f336" Feb 28 10:50:57 crc kubenswrapper[4972]: I0228 10:50:57.295529 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerStarted","Data":"b958fa74140a1f52b3040ba61b3ee98a18c33ed69fcf8091db6758a467bd6b6a"} Feb 28 10:50:57 crc kubenswrapper[4972]: I0228 10:50:57.295595 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerStarted","Data":"aab2a20a96e254981d1aed077b0ae95fcfc6cdc8976aae7ee2ab40e4032a1f1f"} Feb 28 10:50:58 crc kubenswrapper[4972]: I0228 10:50:58.309637 4972 generic.go:334] "Generic (PLEG): container finished" podID="ee147555-f085-4188-89ff-f663afd7ea4e" containerID="b958fa74140a1f52b3040ba61b3ee98a18c33ed69fcf8091db6758a467bd6b6a" exitCode=0 Feb 28 10:50:58 crc kubenswrapper[4972]: I0228 10:50:58.310126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerDied","Data":"b958fa74140a1f52b3040ba61b3ee98a18c33ed69fcf8091db6758a467bd6b6a"} Feb 28 10:50:58 crc kubenswrapper[4972]: I0228 10:50:58.890452 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:50:58 crc kubenswrapper[4972]: I0228 10:50:58.891016 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:50:59 crc kubenswrapper[4972]: I0228 10:50:59.318882 4972 generic.go:334] "Generic (PLEG): container finished" podID="ee147555-f085-4188-89ff-f663afd7ea4e" containerID="a3360a6efbcf2f5ef13cc1a12bb09555fb02c3bede1a6d9ee1b633d92e3c75ad" exitCode=0 Feb 28 10:50:59 crc kubenswrapper[4972]: I0228 10:50:59.318950 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerDied","Data":"a3360a6efbcf2f5ef13cc1a12bb09555fb02c3bede1a6d9ee1b633d92e3c75ad"} Feb 28 10:51:00 crc kubenswrapper[4972]: I0228 10:51:00.327917 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerStarted","Data":"4109d54143653d3f84058e63ec271fcad86d9a0c13ac4ad80f9155ea1bd4887e"} Feb 28 10:51:00 crc kubenswrapper[4972]: I0228 10:51:00.354751 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f6vm2" podStartSLOduration=2.87468017 podStartE2EDuration="4.354722578s" podCreationTimestamp="2026-02-28 10:50:56 +0000 UTC" firstStartedPulling="2026-02-28 10:50:58.311983911 +0000 UTC m=+915.223991649" lastFinishedPulling="2026-02-28 10:50:59.792026329 +0000 UTC m=+916.704034057" observedRunningTime="2026-02-28 10:51:00.352524437 +0000 UTC m=+917.264532165" watchObservedRunningTime="2026-02-28 10:51:00.354722578 +0000 UTC m=+917.266730316" Feb 28 10:51:06 crc kubenswrapper[4972]: I0228 10:51:06.578032 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:06 crc kubenswrapper[4972]: I0228 10:51:06.579335 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:06 crc kubenswrapper[4972]: I0228 10:51:06.695371 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:07 crc kubenswrapper[4972]: I0228 10:51:07.476727 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:07 crc kubenswrapper[4972]: I0228 10:51:07.544260 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.159578 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.161081 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.164916 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-ltpbp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.172048 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.173194 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.178089 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-l829w" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.180120 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.189813 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlc9t\" (UniqueName: \"kubernetes.io/projected/7885ab2e-6366-4513-8904-a982144ac4c4-kube-api-access-mlc9t\") pod \"cinder-operator-controller-manager-768c8b45bb-w987c\" (UID: \"7885ab2e-6366-4513-8904-a982144ac4c4\") " pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.189957 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv2gm\" (UniqueName: \"kubernetes.io/projected/9c739be1-15cd-4044-8cce-c3d2750d030e-kube-api-access-lv2gm\") pod \"barbican-operator-controller-manager-6fb74c6d59-x6hp2\" (UID: \"9c739be1-15cd-4044-8cce-c3d2750d030e\") " pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.194529 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.199823 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.200717 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.203451 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4jjsc" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.224558 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.230814 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.242083 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-frc6q" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.245884 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.249532 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-585b788787-j2pnv"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.250675 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.254902 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-r56kv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.274575 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.281630 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.284823 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.290839 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7rljc" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.292178 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlc9t\" (UniqueName: \"kubernetes.io/projected/7885ab2e-6366-4513-8904-a982144ac4c4-kube-api-access-mlc9t\") pod \"cinder-operator-controller-manager-768c8b45bb-w987c\" (UID: \"7885ab2e-6366-4513-8904-a982144ac4c4\") " pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.292249 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv2gm\" (UniqueName: \"kubernetes.io/projected/9c739be1-15cd-4044-8cce-c3d2750d030e-kube-api-access-lv2gm\") pod \"barbican-operator-controller-manager-6fb74c6d59-x6hp2\" (UID: \"9c739be1-15cd-4044-8cce-c3d2750d030e\") " pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.306874 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-585b788787-j2pnv"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.332974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlc9t\" (UniqueName: \"kubernetes.io/projected/7885ab2e-6366-4513-8904-a982144ac4c4-kube-api-access-mlc9t\") pod \"cinder-operator-controller-manager-768c8b45bb-w987c\" (UID: \"7885ab2e-6366-4513-8904-a982144ac4c4\") " pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.350902 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv2gm\" (UniqueName: \"kubernetes.io/projected/9c739be1-15cd-4044-8cce-c3d2750d030e-kube-api-access-lv2gm\") pod \"barbican-operator-controller-manager-6fb74c6d59-x6hp2\" (UID: \"9c739be1-15cd-4044-8cce-c3d2750d030e\") " pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.357111 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.394716 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkfdx\" (UniqueName: \"kubernetes.io/projected/fb530b5d-27e7-4973-9fb5-41b935e9c0a0-kube-api-access-gkfdx\") pod \"horizon-operator-controller-manager-7db95d7ffb-wdlpp\" (UID: \"fb530b5d-27e7-4973-9fb5-41b935e9c0a0\") " pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.394788 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k9sl\" (UniqueName: \"kubernetes.io/projected/3abb4675-640d-4b5c-b3fb-55279630aecd-kube-api-access-7k9sl\") pod \"heat-operator-controller-manager-585b788787-j2pnv\" (UID: \"3abb4675-640d-4b5c-b3fb-55279630aecd\") " pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.394824 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4h5t\" (UniqueName: \"kubernetes.io/projected/61ba5500-d464-418c-9248-bc05118b0632-kube-api-access-t4h5t\") pod \"designate-operator-controller-manager-55cc45767f-krrsg\" (UID: \"61ba5500-d464-418c-9248-bc05118b0632\") " pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.394851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c97mv\" (UniqueName: \"kubernetes.io/projected/fe59fc2c-7e44-4b09-9d79-415c90f1d3fa-kube-api-access-c97mv\") pod \"glance-operator-controller-manager-5fb5494549-ft8w9\" (UID: \"fe59fc2c-7e44-4b09-9d79-415c90f1d3fa\") " pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.406531 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-c77466965-bkt68"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.407723 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.408293 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.408999 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.415915 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-8vk4x" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.416066 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.416160 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-7xdlp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.433148 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f6vm2" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="registry-server" containerID="cri-o://4109d54143653d3f84058e63ec271fcad86d9a0c13ac4ad80f9155ea1bd4887e" gracePeriod=2 Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.462128 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-c77466965-bkt68"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.481689 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.495908 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.496910 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.498145 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502365 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c97mv\" (UniqueName: \"kubernetes.io/projected/fe59fc2c-7e44-4b09-9d79-415c90f1d3fa-kube-api-access-c97mv\") pod \"glance-operator-controller-manager-5fb5494549-ft8w9\" (UID: \"fe59fc2c-7e44-4b09-9d79-415c90f1d3fa\") " pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502473 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v54sn\" (UniqueName: \"kubernetes.io/projected/20457d28-02f6-4d31-9938-dbdec910d3c8-kube-api-access-v54sn\") pod \"ironic-operator-controller-manager-8784b4656-hhmnw\" (UID: \"20457d28-02f6-4d31-9938-dbdec910d3c8\") " pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502515 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd87k\" (UniqueName: \"kubernetes.io/projected/0c55bc45-0607-477d-893f-5782b00cffb2-kube-api-access-hd87k\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502557 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkfdx\" (UniqueName: \"kubernetes.io/projected/fb530b5d-27e7-4973-9fb5-41b935e9c0a0-kube-api-access-gkfdx\") pod \"horizon-operator-controller-manager-7db95d7ffb-wdlpp\" (UID: \"fb530b5d-27e7-4973-9fb5-41b935e9c0a0\") " pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502602 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502630 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k9sl\" (UniqueName: \"kubernetes.io/projected/3abb4675-640d-4b5c-b3fb-55279630aecd-kube-api-access-7k9sl\") pod \"heat-operator-controller-manager-585b788787-j2pnv\" (UID: \"3abb4675-640d-4b5c-b3fb-55279630aecd\") " pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.502673 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4h5t\" (UniqueName: \"kubernetes.io/projected/61ba5500-d464-418c-9248-bc05118b0632-kube-api-access-t4h5t\") pod \"designate-operator-controller-manager-55cc45767f-krrsg\" (UID: \"61ba5500-d464-418c-9248-bc05118b0632\") " pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.506047 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2wrn8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.520723 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.524424 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkfdx\" (UniqueName: \"kubernetes.io/projected/fb530b5d-27e7-4973-9fb5-41b935e9c0a0-kube-api-access-gkfdx\") pod \"horizon-operator-controller-manager-7db95d7ffb-wdlpp\" (UID: \"fb530b5d-27e7-4973-9fb5-41b935e9c0a0\") " pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.531153 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4h5t\" (UniqueName: \"kubernetes.io/projected/61ba5500-d464-418c-9248-bc05118b0632-kube-api-access-t4h5t\") pod \"designate-operator-controller-manager-55cc45767f-krrsg\" (UID: \"61ba5500-d464-418c-9248-bc05118b0632\") " pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.533408 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.533572 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c97mv\" (UniqueName: \"kubernetes.io/projected/fe59fc2c-7e44-4b09-9d79-415c90f1d3fa-kube-api-access-c97mv\") pod \"glance-operator-controller-manager-5fb5494549-ft8w9\" (UID: \"fe59fc2c-7e44-4b09-9d79-415c90f1d3fa\") " pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.535063 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k9sl\" (UniqueName: \"kubernetes.io/projected/3abb4675-640d-4b5c-b3fb-55279630aecd-kube-api-access-7k9sl\") pod \"heat-operator-controller-manager-585b788787-j2pnv\" (UID: \"3abb4675-640d-4b5c-b3fb-55279630aecd\") " pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.550705 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.569147 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.570514 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.575179 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.601715 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.613511 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-h2mvz" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.613698 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.614642 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.615752 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5gqt\" (UniqueName: \"kubernetes.io/projected/af73edbb-e925-4e12-9cce-58fe2292a497-kube-api-access-s5gqt\") pod \"manila-operator-controller-manager-6f6f57b9b6-rxzp8\" (UID: \"af73edbb-e925-4e12-9cce-58fe2292a497\") " pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.615806 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vq5h\" (UniqueName: \"kubernetes.io/projected/15d3191f-413a-4806-8e55-9dd19db4c335-kube-api-access-7vq5h\") pod \"keystone-operator-controller-manager-78b64779b9-xt2pl\" (UID: \"15d3191f-413a-4806-8e55-9dd19db4c335\") " pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.615840 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v54sn\" (UniqueName: \"kubernetes.io/projected/20457d28-02f6-4d31-9938-dbdec910d3c8-kube-api-access-v54sn\") pod \"ironic-operator-controller-manager-8784b4656-hhmnw\" (UID: \"20457d28-02f6-4d31-9938-dbdec910d3c8\") " pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.615870 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd87k\" (UniqueName: \"kubernetes.io/projected/0c55bc45-0607-477d-893f-5782b00cffb2-kube-api-access-hd87k\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.615906 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: E0228 10:51:09.616034 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:09 crc kubenswrapper[4972]: E0228 10:51:09.616083 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:10.116062313 +0000 UTC m=+927.028070051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.623050 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.634058 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-6kdnm" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.652119 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.653095 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.657241 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v54sn\" (UniqueName: \"kubernetes.io/projected/20457d28-02f6-4d31-9938-dbdec910d3c8-kube-api-access-v54sn\") pod \"ironic-operator-controller-manager-8784b4656-hhmnw\" (UID: \"20457d28-02f6-4d31-9938-dbdec910d3c8\") " pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.674213 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd87k\" (UniqueName: \"kubernetes.io/projected/0c55bc45-0607-477d-893f-5782b00cffb2-kube-api-access-hd87k\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.676865 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-g9p2t" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.723257 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vq5h\" (UniqueName: \"kubernetes.io/projected/15d3191f-413a-4806-8e55-9dd19db4c335-kube-api-access-7vq5h\") pod \"keystone-operator-controller-manager-78b64779b9-xt2pl\" (UID: \"15d3191f-413a-4806-8e55-9dd19db4c335\") " pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.723356 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b685\" (UniqueName: \"kubernetes.io/projected/f2d950d1-acba-46f8-99c7-a932569c1c49-kube-api-access-8b685\") pod \"mariadb-operator-controller-manager-745fc45789-25bfz\" (UID: \"f2d950d1-acba-46f8-99c7-a932569c1c49\") " pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.723428 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkvqf\" (UniqueName: \"kubernetes.io/projected/adaec4ab-a8f3-4719-809d-b102d8e14b7b-kube-api-access-hkvqf\") pod \"nova-operator-controller-manager-6c67ff7674-v8vzt\" (UID: \"adaec4ab-a8f3-4719-809d-b102d8e14b7b\") " pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.723478 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5gqt\" (UniqueName: \"kubernetes.io/projected/af73edbb-e925-4e12-9cce-58fe2292a497-kube-api-access-s5gqt\") pod \"manila-operator-controller-manager-6f6f57b9b6-rxzp8\" (UID: \"af73edbb-e925-4e12-9cce-58fe2292a497\") " pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.748746 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.776183 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.779893 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.784586 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-gbk6h" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.810705 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vq5h\" (UniqueName: \"kubernetes.io/projected/15d3191f-413a-4806-8e55-9dd19db4c335-kube-api-access-7vq5h\") pod \"keystone-operator-controller-manager-78b64779b9-xt2pl\" (UID: \"15d3191f-413a-4806-8e55-9dd19db4c335\") " pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.814369 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5gqt\" (UniqueName: \"kubernetes.io/projected/af73edbb-e925-4e12-9cce-58fe2292a497-kube-api-access-s5gqt\") pod \"manila-operator-controller-manager-6f6f57b9b6-rxzp8\" (UID: \"af73edbb-e925-4e12-9cce-58fe2292a497\") " pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.846736 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkvqf\" (UniqueName: \"kubernetes.io/projected/adaec4ab-a8f3-4719-809d-b102d8e14b7b-kube-api-access-hkvqf\") pod \"nova-operator-controller-manager-6c67ff7674-v8vzt\" (UID: \"adaec4ab-a8f3-4719-809d-b102d8e14b7b\") " pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.846829 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb76f\" (UniqueName: \"kubernetes.io/projected/e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86-kube-api-access-fb76f\") pod \"neutron-operator-controller-manager-768f998cf4-4mxwn\" (UID: \"e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86\") " pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.847079 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b685\" (UniqueName: \"kubernetes.io/projected/f2d950d1-acba-46f8-99c7-a932569c1c49-kube-api-access-8b685\") pod \"mariadb-operator-controller-manager-745fc45789-25bfz\" (UID: \"f2d950d1-acba-46f8-99c7-a932569c1c49\") " pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.884311 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkvqf\" (UniqueName: \"kubernetes.io/projected/adaec4ab-a8f3-4719-809d-b102d8e14b7b-kube-api-access-hkvqf\") pod \"nova-operator-controller-manager-6c67ff7674-v8vzt\" (UID: \"adaec4ab-a8f3-4719-809d-b102d8e14b7b\") " pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.895321 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b685\" (UniqueName: \"kubernetes.io/projected/f2d950d1-acba-46f8-99c7-a932569c1c49-kube-api-access-8b685\") pod \"mariadb-operator-controller-manager-745fc45789-25bfz\" (UID: \"f2d950d1-acba-46f8-99c7-a932569c1c49\") " pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.910365 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.910670 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.911636 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.940878 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.951005 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb76f\" (UniqueName: \"kubernetes.io/projected/e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86-kube-api-access-fb76f\") pod \"neutron-operator-controller-manager-768f998cf4-4mxwn\" (UID: \"e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86\") " pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.976989 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.984306 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb76f\" (UniqueName: \"kubernetes.io/projected/e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86-kube-api-access-fb76f\") pod \"neutron-operator-controller-manager-768f998cf4-4mxwn\" (UID: \"e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86\") " pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.986582 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6"] Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.987717 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.988518 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:09 crc kubenswrapper[4972]: I0228 10:51:09.992857 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gwkd2" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.020580 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.021778 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.025558 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.026659 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.030193 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.030420 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-sd99f" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.030597 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-krgb7" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.054073 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdb26\" (UniqueName: \"kubernetes.io/projected/bae14f0d-c03b-4bf4-9812-2a4cdcae1d94-kube-api-access-vdb26\") pod \"octavia-operator-controller-manager-cc79fdffd-lgrm6\" (UID: \"bae14f0d-c03b-4bf4-9812-2a4cdcae1d94\") " pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.057596 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.065085 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.067964 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.082924 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.111593 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.113859 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.118055 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.118933 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tq9xx" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.127919 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159517 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bmzd\" (UniqueName: \"kubernetes.io/projected/7068aaa9-0410-48a7-9122-53dfad3b2e67-kube-api-access-6bmzd\") pod \"ovn-operator-controller-manager-684c7d77b-x7wnv\" (UID: \"7068aaa9-0410-48a7-9122-53dfad3b2e67\") " pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159579 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzmv\" (UniqueName: \"kubernetes.io/projected/473197b6-6493-4509-920b-ab1be61070f9-kube-api-access-kwzmv\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159605 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159632 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwcx\" (UniqueName: \"kubernetes.io/projected/9c3297f2-e6e9-41e0-b299-fae7573d16cc-kube-api-access-6cwcx\") pod \"placement-operator-controller-manager-bff955cc4-gxsgb\" (UID: \"9c3297f2-e6e9-41e0-b299-fae7573d16cc\") " pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159732 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.159770 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdb26\" (UniqueName: \"kubernetes.io/projected/bae14f0d-c03b-4bf4-9812-2a4cdcae1d94-kube-api-access-vdb26\") pod \"octavia-operator-controller-manager-cc79fdffd-lgrm6\" (UID: \"bae14f0d-c03b-4bf4-9812-2a4cdcae1d94\") " pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.160129 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.160218 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:11.160187346 +0000 UTC m=+928.072195154 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.173354 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.190349 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.191051 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdb26\" (UniqueName: \"kubernetes.io/projected/bae14f0d-c03b-4bf4-9812-2a4cdcae1d94-kube-api-access-vdb26\") pod \"octavia-operator-controller-manager-cc79fdffd-lgrm6\" (UID: \"bae14f0d-c03b-4bf4-9812-2a4cdcae1d94\") " pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.193837 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.204342 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-82nr2" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.210669 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.212141 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.215678 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-pn6jc" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.218042 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.246471 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.256852 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.257857 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.260964 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bmzd\" (UniqueName: \"kubernetes.io/projected/7068aaa9-0410-48a7-9122-53dfad3b2e67-kube-api-access-6bmzd\") pod \"ovn-operator-controller-manager-684c7d77b-x7wnv\" (UID: \"7068aaa9-0410-48a7-9122-53dfad3b2e67\") " pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.261014 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzmv\" (UniqueName: \"kubernetes.io/projected/473197b6-6493-4509-920b-ab1be61070f9-kube-api-access-kwzmv\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.261040 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbj26\" (UniqueName: \"kubernetes.io/projected/510043b9-9bba-49a8-9902-e37f4564bbe8-kube-api-access-sbj26\") pod \"swift-operator-controller-manager-55f4bf89cb-jm2wm\" (UID: \"510043b9-9bba-49a8-9902-e37f4564bbe8\") " pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.261059 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.261080 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwcx\" (UniqueName: \"kubernetes.io/projected/9c3297f2-e6e9-41e0-b299-fae7573d16cc-kube-api-access-6cwcx\") pod \"placement-operator-controller-manager-bff955cc4-gxsgb\" (UID: \"9c3297f2-e6e9-41e0-b299-fae7573d16cc\") " pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.261106 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktck\" (UniqueName: \"kubernetes.io/projected/2da6f8cf-2f39-4614-be52-b6598a919217-kube-api-access-tktck\") pod \"telemetry-operator-controller-manager-56dc67d744-trxff\" (UID: \"2da6f8cf-2f39-4614-be52-b6598a919217\") " pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.261336 4972 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.261384 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert podName:473197b6-6493-4509-920b-ab1be61070f9 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:10.761367842 +0000 UTC m=+927.673375580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" (UID: "473197b6-6493-4509-920b-ab1be61070f9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.262447 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-v94mw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.278536 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.283687 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.284620 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.285967 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bmzd\" (UniqueName: \"kubernetes.io/projected/7068aaa9-0410-48a7-9122-53dfad3b2e67-kube-api-access-6bmzd\") pod \"ovn-operator-controller-manager-684c7d77b-x7wnv\" (UID: \"7068aaa9-0410-48a7-9122-53dfad3b2e67\") " pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.291565 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.294834 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mjt78" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.295207 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzmv\" (UniqueName: \"kubernetes.io/projected/473197b6-6493-4509-920b-ab1be61070f9-kube-api-access-kwzmv\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.303967 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwcx\" (UniqueName: \"kubernetes.io/projected/9c3297f2-e6e9-41e0-b299-fae7573d16cc-kube-api-access-6cwcx\") pod \"placement-operator-controller-manager-bff955cc4-gxsgb\" (UID: \"9c3297f2-e6e9-41e0-b299-fae7573d16cc\") " pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.307659 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.309324 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.312566 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.312688 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-48hfs" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.313037 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.323601 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.332410 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.366145 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.369068 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.376422 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-n9bvp" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.393819 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktck\" (UniqueName: \"kubernetes.io/projected/2da6f8cf-2f39-4614-be52-b6598a919217-kube-api-access-tktck\") pod \"telemetry-operator-controller-manager-56dc67d744-trxff\" (UID: \"2da6f8cf-2f39-4614-be52-b6598a919217\") " pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.393954 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9w9m\" (UniqueName: \"kubernetes.io/projected/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-kube-api-access-x9w9m\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.394060 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.394110 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.394238 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v584\" (UniqueName: \"kubernetes.io/projected/81c08d80-fa23-44c5-9ddf-65dc85b18696-kube-api-access-7v584\") pod \"test-operator-controller-manager-8467ccb4c8-d9mct\" (UID: \"81c08d80-fa23-44c5-9ddf-65dc85b18696\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.394487 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbj26\" (UniqueName: \"kubernetes.io/projected/510043b9-9bba-49a8-9902-e37f4564bbe8-kube-api-access-sbj26\") pod \"swift-operator-controller-manager-55f4bf89cb-jm2wm\" (UID: \"510043b9-9bba-49a8-9902-e37f4564bbe8\") " pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.394541 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq96h\" (UniqueName: \"kubernetes.io/projected/65307cfe-55b4-486c-b154-2196bd0a1012-kube-api-access-kq96h\") pod \"watcher-operator-controller-manager-65c9f4f6b-vpxcf\" (UID: \"65307cfe-55b4-486c-b154-2196bd0a1012\") " pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.408477 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr"] Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.418385 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbj26\" (UniqueName: \"kubernetes.io/projected/510043b9-9bba-49a8-9902-e37f4564bbe8-kube-api-access-sbj26\") pod \"swift-operator-controller-manager-55f4bf89cb-jm2wm\" (UID: \"510043b9-9bba-49a8-9902-e37f4564bbe8\") " pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.444621 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.452153 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktck\" (UniqueName: \"kubernetes.io/projected/2da6f8cf-2f39-4614-be52-b6598a919217-kube-api-access-tktck\") pod \"telemetry-operator-controller-manager-56dc67d744-trxff\" (UID: \"2da6f8cf-2f39-4614-be52-b6598a919217\") " pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.478921 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.493057 4972 generic.go:334] "Generic (PLEG): container finished" podID="ee147555-f085-4188-89ff-f663afd7ea4e" containerID="4109d54143653d3f84058e63ec271fcad86d9a0c13ac4ad80f9155ea1bd4887e" exitCode=0 Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.493105 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerDied","Data":"4109d54143653d3f84058e63ec271fcad86d9a0c13ac4ad80f9155ea1bd4887e"} Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.496820 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v584\" (UniqueName: \"kubernetes.io/projected/81c08d80-fa23-44c5-9ddf-65dc85b18696-kube-api-access-7v584\") pod \"test-operator-controller-manager-8467ccb4c8-d9mct\" (UID: \"81c08d80-fa23-44c5-9ddf-65dc85b18696\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.496905 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq96h\" (UniqueName: \"kubernetes.io/projected/65307cfe-55b4-486c-b154-2196bd0a1012-kube-api-access-kq96h\") pod \"watcher-operator-controller-manager-65c9f4f6b-vpxcf\" (UID: \"65307cfe-55b4-486c-b154-2196bd0a1012\") " pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.496969 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9w9m\" (UniqueName: \"kubernetes.io/projected/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-kube-api-access-x9w9m\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.497017 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.497052 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxmzc\" (UniqueName: \"kubernetes.io/projected/5c4dd163-6c95-4f91-8c40-b242a0d191e3-kube-api-access-hxmzc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8ksfr\" (UID: \"5c4dd163-6c95-4f91-8c40-b242a0d191e3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.497075 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.497250 4972 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.497314 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.497365 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:10.997344129 +0000 UTC m=+927.909351867 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "metrics-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.497397 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:10.99737406 +0000 UTC m=+927.909381798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.513365 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v584\" (UniqueName: \"kubernetes.io/projected/81c08d80-fa23-44c5-9ddf-65dc85b18696-kube-api-access-7v584\") pod \"test-operator-controller-manager-8467ccb4c8-d9mct\" (UID: \"81c08d80-fa23-44c5-9ddf-65dc85b18696\") " pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.515572 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9w9m\" (UniqueName: \"kubernetes.io/projected/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-kube-api-access-x9w9m\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.515876 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq96h\" (UniqueName: \"kubernetes.io/projected/65307cfe-55b4-486c-b154-2196bd0a1012-kube-api-access-kq96h\") pod \"watcher-operator-controller-manager-65c9f4f6b-vpxcf\" (UID: \"65307cfe-55b4-486c-b154-2196bd0a1012\") " pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.589150 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.599203 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxmzc\" (UniqueName: \"kubernetes.io/projected/5c4dd163-6c95-4f91-8c40-b242a0d191e3-kube-api-access-hxmzc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8ksfr\" (UID: \"5c4dd163-6c95-4f91-8c40-b242a0d191e3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.607819 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.620852 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.650280 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxmzc\" (UniqueName: \"kubernetes.io/projected/5c4dd163-6c95-4f91-8c40-b242a0d191e3-kube-api-access-hxmzc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8ksfr\" (UID: \"5c4dd163-6c95-4f91-8c40-b242a0d191e3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.687675 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.715757 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2"] Feb 28 10:51:10 crc kubenswrapper[4972]: W0228 10:51:10.749686 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c739be1_15cd_4044_8cce_c3d2750d030e.slice/crio-28c9d0110e3202a417f6d4e13cf95b15bf83e7056967cf3380606f5a1274887a WatchSource:0}: Error finding container 28c9d0110e3202a417f6d4e13cf95b15bf83e7056967cf3380606f5a1274887a: Status 404 returned error can't find the container with id 28c9d0110e3202a417f6d4e13cf95b15bf83e7056967cf3380606f5a1274887a Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.749902 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.805172 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.805793 4972 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: E0228 10:51:10.805852 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert podName:473197b6-6493-4509-920b-ab1be61070f9 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:11.805826244 +0000 UTC m=+928.717833982 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" (UID: "473197b6-6493-4509-920b-ab1be61070f9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.831720 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.905975 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities\") pod \"ee147555-f085-4188-89ff-f663afd7ea4e\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.906136 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content\") pod \"ee147555-f085-4188-89ff-f663afd7ea4e\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.906197 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxqzz\" (UniqueName: \"kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz\") pod \"ee147555-f085-4188-89ff-f663afd7ea4e\" (UID: \"ee147555-f085-4188-89ff-f663afd7ea4e\") " Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.910382 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities" (OuterVolumeSpecName: "utilities") pod "ee147555-f085-4188-89ff-f663afd7ea4e" (UID: "ee147555-f085-4188-89ff-f663afd7ea4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.926664 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz" (OuterVolumeSpecName: "kube-api-access-zxqzz") pod "ee147555-f085-4188-89ff-f663afd7ea4e" (UID: "ee147555-f085-4188-89ff-f663afd7ea4e"). InnerVolumeSpecName "kube-api-access-zxqzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:51:10 crc kubenswrapper[4972]: I0228 10:51:10.993720 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee147555-f085-4188-89ff-f663afd7ea4e" (UID: "ee147555-f085-4188-89ff-f663afd7ea4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.008361 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.008424 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.008585 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.008606 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee147555-f085-4188-89ff-f663afd7ea4e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.008624 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxqzz\" (UniqueName: \"kubernetes.io/projected/ee147555-f085-4188-89ff-f663afd7ea4e-kube-api-access-zxqzz\") on node \"crc\" DevicePath \"\"" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.008744 4972 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.008773 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.008875 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:12.008804842 +0000 UTC m=+928.920812580 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "metrics-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.008931 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:12.008897505 +0000 UTC m=+928.920905243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.120641 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.124722 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c"] Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.126922 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7885ab2e_6366_4513_8904_a982144ac4c4.slice/crio-36e92117c3539568f614dc31b3c41b3cc3367b1d420bc1c3ad3cae98ca4cd35e WatchSource:0}: Error finding container 36e92117c3539568f614dc31b3c41b3cc3367b1d420bc1c3ad3cae98ca4cd35e: Status 404 returned error can't find the container with id 36e92117c3539568f614dc31b3c41b3cc3367b1d420bc1c3ad3cae98ca4cd35e Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.212824 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.213016 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.213105 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:13.213058676 +0000 UTC m=+930.125066424 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.433884 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt"] Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.455887 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadaec4ab_a8f3_4719_809d_b102d8e14b7b.slice/crio-6eac489eb48304eb6a7891a70a538dee39e3cf7cbe7cf50c507e9b6ed2c59a28 WatchSource:0}: Error finding container 6eac489eb48304eb6a7891a70a538dee39e3cf7cbe7cf50c507e9b6ed2c59a28: Status 404 returned error can't find the container with id 6eac489eb48304eb6a7891a70a538dee39e3cf7cbe7cf50c507e9b6ed2c59a28 Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.469552 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz"] Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.471511 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61ba5500_d464_418c_9248_bc05118b0632.slice/crio-2431ff46300bb0408843150cbe3c6bcc3e13b6c2a829aab6ff00e416f91f2410 WatchSource:0}: Error finding container 2431ff46300bb0408843150cbe3c6bcc3e13b6c2a829aab6ff00e416f91f2410: Status 404 returned error can't find the container with id 2431ff46300bb0408843150cbe3c6bcc3e13b6c2a829aab6ff00e416f91f2410 Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.512674 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.525433 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-585b788787-j2pnv"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.546349 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.553690 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.557120 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" event={"ID":"fb530b5d-27e7-4973-9fb5-41b935e9c0a0","Type":"ContainerStarted","Data":"704b54043b65651de2c11ea33d4f60510504c8a7eaab8a2bc103a4980f8f2a3a"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.560250 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.562002 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6vm2" event={"ID":"ee147555-f085-4188-89ff-f663afd7ea4e","Type":"ContainerDied","Data":"aab2a20a96e254981d1aed077b0ae95fcfc6cdc8976aae7ee2ab40e4032a1f1f"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.562059 4972 scope.go:117] "RemoveContainer" containerID="4109d54143653d3f84058e63ec271fcad86d9a0c13ac4ad80f9155ea1bd4887e" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.562222 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6vm2" Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.566358 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20457d28_02f6_4d31_9938_dbdec910d3c8.slice/crio-b9ed4e38dfa2f9a267483f346bac9916f1b6e0fe4e48361f24b489009d782a5c WatchSource:0}: Error finding container b9ed4e38dfa2f9a267483f346bac9916f1b6e0fe4e48361f24b489009d782a5c: Status 404 returned error can't find the container with id b9ed4e38dfa2f9a267483f346bac9916f1b6e0fe4e48361f24b489009d782a5c Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.570690 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" event={"ID":"adaec4ab-a8f3-4719-809d-b102d8e14b7b","Type":"ContainerStarted","Data":"6eac489eb48304eb6a7891a70a538dee39e3cf7cbe7cf50c507e9b6ed2c59a28"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.578962 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" event={"ID":"61ba5500-d464-418c-9248-bc05118b0632","Type":"ContainerStarted","Data":"2431ff46300bb0408843150cbe3c6bcc3e13b6c2a829aab6ff00e416f91f2410"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.587286 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" event={"ID":"9c739be1-15cd-4044-8cce-c3d2750d030e","Type":"ContainerStarted","Data":"28c9d0110e3202a417f6d4e13cf95b15bf83e7056967cf3380606f5a1274887a"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.590126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" event={"ID":"7885ab2e-6366-4513-8904-a982144ac4c4","Type":"ContainerStarted","Data":"36e92117c3539568f614dc31b3c41b3cc3367b1d420bc1c3ad3cae98ca4cd35e"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.592519 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" event={"ID":"f2d950d1-acba-46f8-99c7-a932569c1c49","Type":"ContainerStarted","Data":"d7f376909305c161e940c56636eb8f0588506efc75bb07d44a977d1490018acc"} Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.606671 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.607691 4972 scope.go:117] "RemoveContainer" containerID="a3360a6efbcf2f5ef13cc1a12bb09555fb02c3bede1a6d9ee1b633d92e3c75ad" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.613190 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f6vm2"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.640679 4972 scope.go:117] "RemoveContainer" containerID="b958fa74140a1f52b3040ba61b3ee98a18c33ed69fcf8091db6758a467bd6b6a" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.736389 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.755489 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv"] Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.757107 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7068aaa9_0410_48a7_9122_53dfad3b2e67.slice/crio-740b2f85b519cbfc24d039eaa5e267d9eb2cd5349f9f7c87ccba333ac62f8a71 WatchSource:0}: Error finding container 740b2f85b519cbfc24d039eaa5e267d9eb2cd5349f9f7c87ccba333ac62f8a71: Status 404 returned error can't find the container with id 740b2f85b519cbfc24d039eaa5e267d9eb2cd5349f9f7c87ccba333ac62f8a71 Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.757801 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81c08d80_fa23_44c5_9ddf_65dc85b18696.slice/crio-97eec8f8e0ad08a73290593def3ea5c294ac1f74bbda92f97fc6d58161073604 WatchSource:0}: Error finding container 97eec8f8e0ad08a73290593def3ea5c294ac1f74bbda92f97fc6d58161073604: Status 404 returned error can't find the container with id 97eec8f8e0ad08a73290593def3ea5c294ac1f74bbda92f97fc6d58161073604 Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.761473 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65307cfe_55b4_486c_b154_2196bd0a1012.slice/crio-296e51fd522042687de5076cbf8f02edfd78f0059355678b3b074ed723c8de60 WatchSource:0}: Error finding container 296e51fd522042687de5076cbf8f02edfd78f0059355678b3b074ed723c8de60: Status 404 returned error can't find the container with id 296e51fd522042687de5076cbf8f02edfd78f0059355678b3b074ed723c8de60 Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.762106 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae14f0d_c03b_4bf4_9812_2a4cdcae1d94.slice/crio-f25ae2db1ab794f236524adf69d7dda3a17c66ede53b2186a08934040d9d7a74 WatchSource:0}: Error finding container f25ae2db1ab794f236524adf69d7dda3a17c66ede53b2186a08934040d9d7a74: Status 404 returned error can't find the container with id f25ae2db1ab794f236524adf69d7dda3a17c66ede53b2186a08934040d9d7a74 Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.765354 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm"] Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.770602 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hxmzc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-8ksfr_openstack-operators(5c4dd163-6c95-4f91-8c40-b242a0d191e3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.772781 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" podUID="5c4dd163-6c95-4f91-8c40-b242a0d191e3" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.773839 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9a940ee50452c206923805ba7bf69dded7fcf53cb7ec14e22e793bd56501e242,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kq96h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-65c9f4f6b-vpxcf_openstack-operators(65307cfe-55b4-486c-b154-2196bd0a1012): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.775699 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" podUID="65307cfe-55b4-486c-b154-2196bd0a1012" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.787436 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6"] Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.799201 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf73edbb_e925_4e12_9cce_58fe2292a497.slice/crio-57587e97c34c178101cb03340cd6b902d9f24957bcd183c2ad0e097fd2ddf16e WatchSource:0}: Error finding container 57587e97c34c178101cb03340cd6b902d9f24957bcd183c2ad0e097fd2ddf16e: Status 404 returned error can't find the container with id 57587e97c34c178101cb03340cd6b902d9f24957bcd183c2ad0e097fd2ddf16e Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.802361 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:6e988fa8bacb3367dea2e02d28abf23403affdb604ca0353473264ec21051ff2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s5gqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6f6f57b9b6-rxzp8_openstack-operators(af73edbb-e925-4e12-9cce-58fe2292a497): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.803517 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" podUID="af73edbb-e925-4e12-9cce-58fe2292a497" Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.803747 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c3297f2_e6e9_41e0_b299_fae7573d16cc.slice/crio-685a33ffbb1ab52c486f57471e4c8a3ddb967de36c839fc1466c6e3b58808db4 WatchSource:0}: Error finding container 685a33ffbb1ab52c486f57471e4c8a3ddb967de36c839fc1466c6e3b58808db4: Status 404 returned error can't find the container with id 685a33ffbb1ab52c486f57471e4c8a3ddb967de36c839fc1466c6e3b58808db4 Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.810549 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" path="/var/lib/kubelet/pods/ee147555-f085-4188-89ff-f663afd7ea4e/volumes" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.811165 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.811194 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr"] Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.811582 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:b61730aa07404c6893c94c73cb7c80f16eb4d92a759740393430aca41f416b28,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cwcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-bff955cc4-gxsgb_openstack-operators(9c3297f2-e6e9-41e0-b299-fae7573d16cc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.812895 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" podUID="9c3297f2-e6e9-41e0-b299-fae7573d16cc" Feb 28 10:51:11 crc kubenswrapper[4972]: W0228 10:51:11.813873 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode099f5ae_f38b_4cb6_8be1_bbd8f91c7e86.slice/crio-128af5889805770cdb5f4dc7fce15cd3043ea6556dda50739be9406ad551f3a4 WatchSource:0}: Error finding container 128af5889805770cdb5f4dc7fce15cd3043ea6556dda50739be9406ad551f3a4: Status 404 returned error can't find the container with id 128af5889805770cdb5f4dc7fce15cd3043ea6556dda50739be9406ad551f3a4 Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.816916 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:d2b850bc2ec026f8a179d5f59ad65b79f2d329e91a4ec8f140a645ebc38069b6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fb76f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-768f998cf4-4mxwn_openstack-operators(e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.818207 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8"] Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.818254 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" podUID="e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.826369 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.828359 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.828543 4972 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.828590 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert podName:473197b6-6493-4509-920b-ab1be61070f9 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:13.828575216 +0000 UTC m=+930.740582954 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" (UID: "473197b6-6493-4509-920b-ab1be61070f9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.842235 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:4b10e23983c3ec518c35aeabb33ac228063e56c81b4d7a100c5d91139ad7d7fc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tktck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-56dc67d744-trxff_openstack-operators(2da6f8cf-2f39-4614-be52-b6598a919217): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 28 10:51:11 crc kubenswrapper[4972]: E0228 10:51:11.844065 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" podUID="2da6f8cf-2f39-4614-be52-b6598a919217" Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.849906 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb"] Feb 28 10:51:11 crc kubenswrapper[4972]: I0228 10:51:11.858675 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff"] Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.032496 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.032600 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:14.032573903 +0000 UTC m=+930.944581781 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.032575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.032709 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.032978 4972 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.033066 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:14.033045116 +0000 UTC m=+930.945052854 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "metrics-server-cert" not found Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.609093 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" event={"ID":"fe59fc2c-7e44-4b09-9d79-415c90f1d3fa","Type":"ContainerStarted","Data":"5eaa438db6411396b63ae04cde0b701f34b603c868dc4c31c1cca57505532595"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.610998 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" event={"ID":"af73edbb-e925-4e12-9cce-58fe2292a497","Type":"ContainerStarted","Data":"57587e97c34c178101cb03340cd6b902d9f24957bcd183c2ad0e097fd2ddf16e"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.612278 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" event={"ID":"15d3191f-413a-4806-8e55-9dd19db4c335","Type":"ContainerStarted","Data":"52b7c88f1495933bb09ba249cddc05a335e573fda4fdc2d7fb33593e0ac7f9f7"} Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.613442 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:6e988fa8bacb3367dea2e02d28abf23403affdb604ca0353473264ec21051ff2\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" podUID="af73edbb-e925-4e12-9cce-58fe2292a497" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.615593 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" event={"ID":"65307cfe-55b4-486c-b154-2196bd0a1012","Type":"ContainerStarted","Data":"296e51fd522042687de5076cbf8f02edfd78f0059355678b3b074ed723c8de60"} Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.617153 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9a940ee50452c206923805ba7bf69dded7fcf53cb7ec14e22e793bd56501e242\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" podUID="65307cfe-55b4-486c-b154-2196bd0a1012" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.617646 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" event={"ID":"81c08d80-fa23-44c5-9ddf-65dc85b18696","Type":"ContainerStarted","Data":"97eec8f8e0ad08a73290593def3ea5c294ac1f74bbda92f97fc6d58161073604"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.619469 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" event={"ID":"5c4dd163-6c95-4f91-8c40-b242a0d191e3","Type":"ContainerStarted","Data":"9b87ef2db6d2289aa7279df410dde8fe2442c6e53cd698af64ef270190c4715a"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.620519 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" event={"ID":"e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86","Type":"ContainerStarted","Data":"128af5889805770cdb5f4dc7fce15cd3043ea6556dda50739be9406ad551f3a4"} Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.621064 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" podUID="5c4dd163-6c95-4f91-8c40-b242a0d191e3" Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.621241 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:d2b850bc2ec026f8a179d5f59ad65b79f2d329e91a4ec8f140a645ebc38069b6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" podUID="e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.637794 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" event={"ID":"9c3297f2-e6e9-41e0-b299-fae7573d16cc","Type":"ContainerStarted","Data":"685a33ffbb1ab52c486f57471e4c8a3ddb967de36c839fc1466c6e3b58808db4"} Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.639609 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:b61730aa07404c6893c94c73cb7c80f16eb4d92a759740393430aca41f416b28\\\"\"" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" podUID="9c3297f2-e6e9-41e0-b299-fae7573d16cc" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.640274 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" event={"ID":"2da6f8cf-2f39-4614-be52-b6598a919217","Type":"ContainerStarted","Data":"e50bf73e633569b21feda08cd6d8e27a236afffdc94e5987d63104f9cb9ae955"} Feb 28 10:51:12 crc kubenswrapper[4972]: E0228 10:51:12.641553 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:4b10e23983c3ec518c35aeabb33ac228063e56c81b4d7a100c5d91139ad7d7fc\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" podUID="2da6f8cf-2f39-4614-be52-b6598a919217" Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.644797 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" event={"ID":"7068aaa9-0410-48a7-9122-53dfad3b2e67","Type":"ContainerStarted","Data":"740b2f85b519cbfc24d039eaa5e267d9eb2cd5349f9f7c87ccba333ac62f8a71"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.646370 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" event={"ID":"510043b9-9bba-49a8-9902-e37f4564bbe8","Type":"ContainerStarted","Data":"cab2cdce26d19ea1b69192f4cabc4b3bbc41bbb406d2feb38d2624a4287601aa"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.650115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" event={"ID":"bae14f0d-c03b-4bf4-9812-2a4cdcae1d94","Type":"ContainerStarted","Data":"f25ae2db1ab794f236524adf69d7dda3a17c66ede53b2186a08934040d9d7a74"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.651719 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" event={"ID":"3abb4675-640d-4b5c-b3fb-55279630aecd","Type":"ContainerStarted","Data":"8af5590cd0e0d21ab214e178c8bd8797dabfb9c7d62e48fd91ea559e88f6d470"} Feb 28 10:51:12 crc kubenswrapper[4972]: I0228 10:51:12.658294 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" event={"ID":"20457d28-02f6-4d31-9938-dbdec910d3c8","Type":"ContainerStarted","Data":"b9ed4e38dfa2f9a267483f346bac9916f1b6e0fe4e48361f24b489009d782a5c"} Feb 28 10:51:13 crc kubenswrapper[4972]: I0228 10:51:13.257189 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.258008 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.258087 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:17.258049477 +0000 UTC m=+934.170057215 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.684197 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:d2b850bc2ec026f8a179d5f59ad65b79f2d329e91a4ec8f140a645ebc38069b6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" podUID="e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.685256 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9a940ee50452c206923805ba7bf69dded7fcf53cb7ec14e22e793bd56501e242\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" podUID="65307cfe-55b4-486c-b154-2196bd0a1012" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.685315 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:4b10e23983c3ec518c35aeabb33ac228063e56c81b4d7a100c5d91139ad7d7fc\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" podUID="2da6f8cf-2f39-4614-be52-b6598a919217" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.685356 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:b61730aa07404c6893c94c73cb7c80f16eb4d92a759740393430aca41f416b28\\\"\"" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" podUID="9c3297f2-e6e9-41e0-b299-fae7573d16cc" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.685432 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:6e988fa8bacb3367dea2e02d28abf23403affdb604ca0353473264ec21051ff2\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" podUID="af73edbb-e925-4e12-9cce-58fe2292a497" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.686059 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" podUID="5c4dd163-6c95-4f91-8c40-b242a0d191e3" Feb 28 10:51:13 crc kubenswrapper[4972]: I0228 10:51:13.867908 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.868218 4972 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:13 crc kubenswrapper[4972]: E0228 10:51:13.868301 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert podName:473197b6-6493-4509-920b-ab1be61070f9 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:17.868274679 +0000 UTC m=+934.780282417 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" (UID: "473197b6-6493-4509-920b-ab1be61070f9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:14 crc kubenswrapper[4972]: I0228 10:51:14.072515 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:14 crc kubenswrapper[4972]: I0228 10:51:14.072594 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:14 crc kubenswrapper[4972]: E0228 10:51:14.072778 4972 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 28 10:51:14 crc kubenswrapper[4972]: E0228 10:51:14.072844 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:18.072823901 +0000 UTC m=+934.984831639 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "metrics-server-cert" not found Feb 28 10:51:14 crc kubenswrapper[4972]: E0228 10:51:14.072873 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:14 crc kubenswrapper[4972]: E0228 10:51:14.072941 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:18.072920544 +0000 UTC m=+934.984928272 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:17 crc kubenswrapper[4972]: I0228 10:51:17.340275 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:17 crc kubenswrapper[4972]: E0228 10:51:17.340548 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:17 crc kubenswrapper[4972]: E0228 10:51:17.340895 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:25.340870319 +0000 UTC m=+942.252878057 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:17 crc kubenswrapper[4972]: I0228 10:51:17.949597 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:17 crc kubenswrapper[4972]: E0228 10:51:17.949857 4972 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:17 crc kubenswrapper[4972]: E0228 10:51:17.949964 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert podName:473197b6-6493-4509-920b-ab1be61070f9 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:25.949937398 +0000 UTC m=+942.861945146 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert") pod "openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" (UID: "473197b6-6493-4509-920b-ab1be61070f9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 28 10:51:18 crc kubenswrapper[4972]: I0228 10:51:18.153859 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:18 crc kubenswrapper[4972]: I0228 10:51:18.153946 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:18 crc kubenswrapper[4972]: E0228 10:51:18.154087 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:18 crc kubenswrapper[4972]: E0228 10:51:18.154178 4972 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 28 10:51:18 crc kubenswrapper[4972]: E0228 10:51:18.154181 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:26.154161211 +0000 UTC m=+943.066168949 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:18 crc kubenswrapper[4972]: E0228 10:51:18.154278 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:26.154257614 +0000 UTC m=+943.066265522 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "metrics-server-cert" not found Feb 28 10:51:24 crc kubenswrapper[4972]: E0228 10:51:24.938966 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:b235599fe44c901b7ac0b51dfbcc9e0cea2bf5a9dc8295bafe16bba528d72997" Feb 28 10:51:24 crc kubenswrapper[4972]: E0228 10:51:24.939659 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:b235599fe44c901b7ac0b51dfbcc9e0cea2bf5a9dc8295bafe16bba528d72997,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7vq5h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-78b64779b9-xt2pl_openstack-operators(15d3191f-413a-4806-8e55-9dd19db4c335): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:51:24 crc kubenswrapper[4972]: E0228 10:51:24.940874 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" podUID="15d3191f-413a-4806-8e55-9dd19db4c335" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.398425 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:25 crc kubenswrapper[4972]: E0228 10:51:25.398705 4972 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:25 crc kubenswrapper[4972]: E0228 10:51:25.398927 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert podName:0c55bc45-0607-477d-893f-5782b00cffb2 nodeName:}" failed. No retries permitted until 2026-02-28 10:51:41.398905687 +0000 UTC m=+958.310913415 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert") pod "infra-operator-controller-manager-c77466965-bkt68" (UID: "0c55bc45-0607-477d-893f-5782b00cffb2") : secret "infra-operator-webhook-server-cert" not found Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.803100 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" event={"ID":"510043b9-9bba-49a8-9902-e37f4564bbe8","Type":"ContainerStarted","Data":"a0486a8c8b4df1027b03ca41df971bf14d536a50a03dce3723375bb499655150"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.803183 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.814837 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" event={"ID":"adaec4ab-a8f3-4719-809d-b102d8e14b7b","Type":"ContainerStarted","Data":"17808a16369e9e1359757c12cd76b30025bab979bec6c7893d55c631eb455af6"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.815020 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.822018 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" event={"ID":"f2d950d1-acba-46f8-99c7-a932569c1c49","Type":"ContainerStarted","Data":"b5747536a2d0dadce7a2789fbcc2d896d526107a43a1f215ba2df227d3a0ac6e"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.822448 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.828135 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" event={"ID":"3abb4675-640d-4b5c-b3fb-55279630aecd","Type":"ContainerStarted","Data":"b3f9227b662604621ba0f6a68695d9c2282a3efcb785e23390b721f92402a4dd"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.829064 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.838755 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" event={"ID":"61ba5500-d464-418c-9248-bc05118b0632","Type":"ContainerStarted","Data":"c3f7d44dd7b8558c4463274ebf700602ecbb3d2da8d8f863fd2733f65a043dea"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.839676 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.863946 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" event={"ID":"20457d28-02f6-4d31-9938-dbdec910d3c8","Type":"ContainerStarted","Data":"9baa3b3ecc8409e222a5e3bf56bf911706c4801f99a27a27c53651e02dac2589"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.864995 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.875850 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" event={"ID":"fe59fc2c-7e44-4b09-9d79-415c90f1d3fa","Type":"ContainerStarted","Data":"8cc5c44ee83792d59ba5b67131b8429faf65e3bf165c1764c4e789c1466a2079"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.876712 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.881693 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" event={"ID":"7885ab2e-6366-4513-8904-a982144ac4c4","Type":"ContainerStarted","Data":"2f2f4a01ade974a400e85466d85dabfb06e1458a6026b61ca44e58fbec5a4ecc"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.882169 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.893532 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" event={"ID":"bae14f0d-c03b-4bf4-9812-2a4cdcae1d94","Type":"ContainerStarted","Data":"1f952d45f440d4ee7e5e35af540655a47f788ab3d8a523a8e87eac0373a6ce44"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.894180 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.895423 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" event={"ID":"7068aaa9-0410-48a7-9122-53dfad3b2e67","Type":"ContainerStarted","Data":"3e92409c44b99d74ded8e240e1a87b93bab8f3f5a490cc0f363df595b546a8e7"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.895835 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.896779 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" event={"ID":"fb530b5d-27e7-4973-9fb5-41b935e9c0a0","Type":"ContainerStarted","Data":"ed81ac3cfa769201fd14d1e8b035d3876e63a8346396cdaa711ec6b3928807b5"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.897169 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.925535 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" event={"ID":"81c08d80-fa23-44c5-9ddf-65dc85b18696","Type":"ContainerStarted","Data":"2d1d5f44c122db9b71aa51312021b7d98bca5c4c432bbe175d0c5f38133a7cc1"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.926291 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.940155 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" event={"ID":"9c739be1-15cd-4044-8cce-c3d2750d030e","Type":"ContainerStarted","Data":"3fdeb4481fb4ec65dfb4a1508f5f34bcf1a2f04c77ddc379e9308d7a6ea29e36"} Feb 28 10:51:25 crc kubenswrapper[4972]: I0228 10:51:25.940210 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:25 crc kubenswrapper[4972]: E0228 10:51:25.943719 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:b235599fe44c901b7ac0b51dfbcc9e0cea2bf5a9dc8295bafe16bba528d72997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" podUID="15d3191f-413a-4806-8e55-9dd19db4c335" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.003677 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" podStartSLOduration=3.814567195 podStartE2EDuration="17.003659927s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.7636862 +0000 UTC m=+928.675693938" lastFinishedPulling="2026-02-28 10:51:24.952778932 +0000 UTC m=+941.864786670" observedRunningTime="2026-02-28 10:51:25.924956347 +0000 UTC m=+942.836964085" watchObservedRunningTime="2026-02-28 10:51:26.003659927 +0000 UTC m=+942.915667665" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.006417 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" podStartSLOduration=3.543027079 podStartE2EDuration="17.006408473s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.447158282 +0000 UTC m=+928.359166020" lastFinishedPulling="2026-02-28 10:51:24.910539666 +0000 UTC m=+941.822547414" observedRunningTime="2026-02-28 10:51:25.969091735 +0000 UTC m=+942.881099473" watchObservedRunningTime="2026-02-28 10:51:26.006408473 +0000 UTC m=+942.918416211" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.008281 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.017210 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/473197b6-6493-4509-920b-ab1be61070f9-cert\") pod \"openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw\" (UID: \"473197b6-6493-4509-920b-ab1be61070f9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.045488 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" podStartSLOduration=3.717836304 podStartE2EDuration="17.04544923s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.582084946 +0000 UTC m=+928.494092684" lastFinishedPulling="2026-02-28 10:51:24.909697872 +0000 UTC m=+941.821705610" observedRunningTime="2026-02-28 10:51:26.03609576 +0000 UTC m=+942.948103498" watchObservedRunningTime="2026-02-28 10:51:26.04544923 +0000 UTC m=+942.957456968" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.132187 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" podStartSLOduration=3.652996579 podStartE2EDuration="17.132167003s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.473589167 +0000 UTC m=+928.385596905" lastFinishedPulling="2026-02-28 10:51:24.952759591 +0000 UTC m=+941.864767329" observedRunningTime="2026-02-28 10:51:26.119401437 +0000 UTC m=+943.031409175" watchObservedRunningTime="2026-02-28 10:51:26.132167003 +0000 UTC m=+943.044174741" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.215785 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.216197 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:26 crc kubenswrapper[4972]: E0228 10:51:26.217718 4972 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 28 10:51:26 crc kubenswrapper[4972]: E0228 10:51:26.217824 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs podName:8cce4f5d-78aa-44c4-8c60-8bce681ada4e nodeName:}" failed. No retries permitted until 2026-02-28 10:51:42.217800786 +0000 UTC m=+959.129808524 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs") pod "openstack-operator-controller-manager-6cf56d77c9-f7stz" (UID: "8cce4f5d-78aa-44c4-8c60-8bce681ada4e") : secret "webhook-server-cert" not found Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.248247 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-metrics-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.279107 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.368992 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" podStartSLOduration=4.225751439 podStartE2EDuration="17.368972903s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.766446727 +0000 UTC m=+928.678454465" lastFinishedPulling="2026-02-28 10:51:24.909668191 +0000 UTC m=+941.821675929" observedRunningTime="2026-02-28 10:51:26.368487749 +0000 UTC m=+943.280495487" watchObservedRunningTime="2026-02-28 10:51:26.368972903 +0000 UTC m=+943.280980641" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.371240 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" podStartSLOduration=4.180230151 podStartE2EDuration="17.371233576s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.761654593 +0000 UTC m=+928.673662331" lastFinishedPulling="2026-02-28 10:51:24.952658018 +0000 UTC m=+941.864665756" observedRunningTime="2026-02-28 10:51:26.272387735 +0000 UTC m=+943.184395463" watchObservedRunningTime="2026-02-28 10:51:26.371233576 +0000 UTC m=+943.283241314" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.437237 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" podStartSLOduration=3.656829626 podStartE2EDuration="17.437217203s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.130024496 +0000 UTC m=+928.042032234" lastFinishedPulling="2026-02-28 10:51:24.910412073 +0000 UTC m=+941.822419811" observedRunningTime="2026-02-28 10:51:26.429855197 +0000 UTC m=+943.341862935" watchObservedRunningTime="2026-02-28 10:51:26.437217203 +0000 UTC m=+943.349224951" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.494601 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" podStartSLOduration=4.125310942 podStartE2EDuration="17.494583818s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.541088925 +0000 UTC m=+928.453096663" lastFinishedPulling="2026-02-28 10:51:24.910361811 +0000 UTC m=+941.822369539" observedRunningTime="2026-02-28 10:51:26.492946663 +0000 UTC m=+943.404954401" watchObservedRunningTime="2026-02-28 10:51:26.494583818 +0000 UTC m=+943.406591546" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.495733 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" podStartSLOduration=3.98758416 podStartE2EDuration="17.495728481s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.465285445 +0000 UTC m=+928.377293183" lastFinishedPulling="2026-02-28 10:51:24.973429746 +0000 UTC m=+941.885437504" observedRunningTime="2026-02-28 10:51:26.468924185 +0000 UTC m=+943.380931923" watchObservedRunningTime="2026-02-28 10:51:26.495728481 +0000 UTC m=+943.407736219" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.551254 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" podStartSLOduration=4.157747105 podStartE2EDuration="17.551238735s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.558662265 +0000 UTC m=+928.470670003" lastFinishedPulling="2026-02-28 10:51:24.952153895 +0000 UTC m=+941.864161633" observedRunningTime="2026-02-28 10:51:26.546264157 +0000 UTC m=+943.458271895" watchObservedRunningTime="2026-02-28 10:51:26.551238735 +0000 UTC m=+943.463246473" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.615054 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" podStartSLOduration=3.834247354 podStartE2EDuration="17.615033471s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.128595456 +0000 UTC m=+928.040603194" lastFinishedPulling="2026-02-28 10:51:24.909381573 +0000 UTC m=+941.821389311" observedRunningTime="2026-02-28 10:51:26.57725791 +0000 UTC m=+943.489265638" watchObservedRunningTime="2026-02-28 10:51:26.615033471 +0000 UTC m=+943.527041209" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.616132 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" podStartSLOduration=3.479024188 podStartE2EDuration="17.616125112s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:10.773509454 +0000 UTC m=+927.685517182" lastFinishedPulling="2026-02-28 10:51:24.910610368 +0000 UTC m=+941.822618106" observedRunningTime="2026-02-28 10:51:26.610096443 +0000 UTC m=+943.522104181" watchObservedRunningTime="2026-02-28 10:51:26.616125112 +0000 UTC m=+943.528132850" Feb 28 10:51:26 crc kubenswrapper[4972]: I0228 10:51:26.702149 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" podStartSLOduration=4.560003169 podStartE2EDuration="17.702130814s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.768312318 +0000 UTC m=+928.680320056" lastFinishedPulling="2026-02-28 10:51:24.910439963 +0000 UTC m=+941.822447701" observedRunningTime="2026-02-28 10:51:26.666073221 +0000 UTC m=+943.578080959" watchObservedRunningTime="2026-02-28 10:51:26.702130814 +0000 UTC m=+943.614138552" Feb 28 10:51:27 crc kubenswrapper[4972]: I0228 10:51:27.111156 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw"] Feb 28 10:51:27 crc kubenswrapper[4972]: W0228 10:51:27.141418 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod473197b6_6493_4509_920b_ab1be61070f9.slice/crio-b8ee6d2392e822f560c23450853064f81d84abd8bb8a87e21291462053cc44ac WatchSource:0}: Error finding container b8ee6d2392e822f560c23450853064f81d84abd8bb8a87e21291462053cc44ac: Status 404 returned error can't find the container with id b8ee6d2392e822f560c23450853064f81d84abd8bb8a87e21291462053cc44ac Feb 28 10:51:27 crc kubenswrapper[4972]: I0228 10:51:27.962570 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" event={"ID":"473197b6-6493-4509-920b-ab1be61070f9","Type":"ContainerStarted","Data":"b8ee6d2392e822f560c23450853064f81d84abd8bb8a87e21291462053cc44ac"} Feb 28 10:51:28 crc kubenswrapper[4972]: I0228 10:51:28.891179 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:51:28 crc kubenswrapper[4972]: I0228 10:51:28.891254 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:51:28 crc kubenswrapper[4972]: I0228 10:51:28.891317 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:51:28 crc kubenswrapper[4972]: I0228 10:51:28.892110 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:51:28 crc kubenswrapper[4972]: I0228 10:51:28.892164 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3" gracePeriod=600 Feb 28 10:51:29 crc kubenswrapper[4972]: I0228 10:51:29.988550 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3" exitCode=0 Feb 28 10:51:29 crc kubenswrapper[4972]: I0228 10:51:29.988654 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3"} Feb 28 10:51:29 crc kubenswrapper[4972]: I0228 10:51:29.988966 4972 scope.go:117] "RemoveContainer" containerID="381019b90d378b5091027b9ec7568709ea945442da72c493864a9dfe040c7298" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.074565 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c67ff7674-v8vzt" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.087758 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-745fc45789-25bfz" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.344115 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-cc79fdffd-lgrm6" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.451032 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-684c7d77b-x7wnv" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.592846 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8467ccb4c8-d9mct" Feb 28 10:51:30 crc kubenswrapper[4972]: I0228 10:51:30.624360 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.053648 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" event={"ID":"473197b6-6493-4509-920b-ab1be61070f9","Type":"ContainerStarted","Data":"2fecf41b836eb3182427c86bc6b550e2d5ab976df2f5f035a87dc40bcfff052b"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.055553 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.056121 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" event={"ID":"e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86","Type":"ContainerStarted","Data":"a34522838e8bd70787c4106e8b2e61c6635defe5f85afde1aaf465945ad40762"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.056330 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.057650 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" event={"ID":"9c3297f2-e6e9-41e0-b299-fae7573d16cc","Type":"ContainerStarted","Data":"a5168e763b4c21e3ddd407abcfc9299220ed2b9e112ca3311050bded97d2b09c"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.057791 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.059815 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" event={"ID":"2da6f8cf-2f39-4614-be52-b6598a919217","Type":"ContainerStarted","Data":"d39f45a2cb8db6a755cc3aaea8af6d05bc2c5bc5075a0cac14bea73bf3ed9af8"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.060013 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.061311 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" event={"ID":"af73edbb-e925-4e12-9cce-58fe2292a497","Type":"ContainerStarted","Data":"7c4bd9da3855266049ae44f0669cd0683efc36b70d73f260c03384e13ff450ef"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.061487 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.062789 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" event={"ID":"65307cfe-55b4-486c-b154-2196bd0a1012","Type":"ContainerStarted","Data":"1c997e13680fbdfbe1e13db57affaa7e6524e14e74e018ef17b6afef19135978"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.062967 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.064963 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.066822 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" event={"ID":"5c4dd163-6c95-4f91-8c40-b242a0d191e3","Type":"ContainerStarted","Data":"f94dbef4e5682ecc05d8178a8f9b868af99016d1525b6743aeb105004dbc9994"} Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.083567 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" podStartSLOduration=19.445816488 podStartE2EDuration="26.083541372s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:27.153498935 +0000 UTC m=+944.065506673" lastFinishedPulling="2026-02-28 10:51:33.791223819 +0000 UTC m=+950.703231557" observedRunningTime="2026-02-28 10:51:35.077320659 +0000 UTC m=+951.989328397" watchObservedRunningTime="2026-02-28 10:51:35.083541372 +0000 UTC m=+951.995549110" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.095561 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" podStartSLOduration=4.106449667 podStartE2EDuration="26.095537416s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.801514762 +0000 UTC m=+928.713522500" lastFinishedPulling="2026-02-28 10:51:33.790602511 +0000 UTC m=+950.702610249" observedRunningTime="2026-02-28 10:51:35.093281004 +0000 UTC m=+952.005288742" watchObservedRunningTime="2026-02-28 10:51:35.095537416 +0000 UTC m=+952.007545154" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.117546 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" podStartSLOduration=4.100725258 podStartE2EDuration="26.117523588s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.773708079 +0000 UTC m=+928.685715817" lastFinishedPulling="2026-02-28 10:51:33.790506409 +0000 UTC m=+950.702514147" observedRunningTime="2026-02-28 10:51:35.115530053 +0000 UTC m=+952.027537791" watchObservedRunningTime="2026-02-28 10:51:35.117523588 +0000 UTC m=+952.029531326" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.157822 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" podStartSLOduration=4.212487749 podStartE2EDuration="26.157799239s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.816799028 +0000 UTC m=+928.728806766" lastFinishedPulling="2026-02-28 10:51:33.762110518 +0000 UTC m=+950.674118256" observedRunningTime="2026-02-28 10:51:35.152341987 +0000 UTC m=+952.064349745" watchObservedRunningTime="2026-02-28 10:51:35.157799239 +0000 UTC m=+952.069806977" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.174668 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" podStartSLOduration=4.225398729 podStartE2EDuration="26.174638798s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.811361187 +0000 UTC m=+928.723368925" lastFinishedPulling="2026-02-28 10:51:33.760601256 +0000 UTC m=+950.672608994" observedRunningTime="2026-02-28 10:51:35.171331296 +0000 UTC m=+952.083339034" watchObservedRunningTime="2026-02-28 10:51:35.174638798 +0000 UTC m=+952.086646536" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.227963 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8ksfr" podStartSLOduration=3.098426925 podStartE2EDuration="25.227929421s" podCreationTimestamp="2026-02-28 10:51:10 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.770404797 +0000 UTC m=+928.682412535" lastFinishedPulling="2026-02-28 10:51:33.899907283 +0000 UTC m=+950.811915031" observedRunningTime="2026-02-28 10:51:35.209176219 +0000 UTC m=+952.121183977" watchObservedRunningTime="2026-02-28 10:51:35.227929421 +0000 UTC m=+952.139937159" Feb 28 10:51:35 crc kubenswrapper[4972]: I0228 10:51:35.249106 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" podStartSLOduration=4.330674848 podStartE2EDuration="26.249080879s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.841978999 +0000 UTC m=+928.753986737" lastFinishedPulling="2026-02-28 10:51:33.76038503 +0000 UTC m=+950.672392768" observedRunningTime="2026-02-28 10:51:35.245785267 +0000 UTC m=+952.157793025" watchObservedRunningTime="2026-02-28 10:51:35.249080879 +0000 UTC m=+952.161088617" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.503001 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6fb74c6d59-x6hp2" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.530596 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-768c8b45bb-w987c" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.537999 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-55cc45767f-krrsg" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.578189 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5fb5494549-ft8w9" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.608580 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-585b788787-j2pnv" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.639099 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7db95d7ffb-wdlpp" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.752530 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-8784b4656-hhmnw" Feb 28 10:51:39 crc kubenswrapper[4972]: I0228 10:51:39.993084 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6f6f57b9b6-rxzp8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.124663 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-768f998cf4-4mxwn" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.482809 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-bff955cc4-gxsgb" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.586180 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:51:40 crc kubenswrapper[4972]: E0228 10:51:40.587029 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="extract-utilities" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.587053 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="extract-utilities" Feb 28 10:51:40 crc kubenswrapper[4972]: E0228 10:51:40.587081 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="extract-content" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.587095 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="extract-content" Feb 28 10:51:40 crc kubenswrapper[4972]: E0228 10:51:40.587112 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="registry-server" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.587123 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="registry-server" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.587344 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee147555-f085-4188-89ff-f663afd7ea4e" containerName="registry-server" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.588573 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.599118 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.611686 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-65c9f4f6b-vpxcf" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.742777 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.742942 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d484f\" (UniqueName: \"kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.742992 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.754151 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-56dc67d744-trxff" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.845250 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.845389 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d484f\" (UniqueName: \"kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.845438 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.845832 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.846101 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.875760 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d484f\" (UniqueName: \"kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f\") pod \"community-operators-24lg8\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:40 crc kubenswrapper[4972]: I0228 10:51:40.934385 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:41 crc kubenswrapper[4972]: I0228 10:51:41.281281 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:51:41 crc kubenswrapper[4972]: I0228 10:51:41.461994 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:41 crc kubenswrapper[4972]: I0228 10:51:41.469258 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c55bc45-0607-477d-893f-5782b00cffb2-cert\") pod \"infra-operator-controller-manager-c77466965-bkt68\" (UID: \"0c55bc45-0607-477d-893f-5782b00cffb2\") " pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:41 crc kubenswrapper[4972]: I0228 10:51:41.562888 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.085646 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-c77466965-bkt68"] Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.131653 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" event={"ID":"0c55bc45-0607-477d-893f-5782b00cffb2","Type":"ContainerStarted","Data":"e7ed3abab2471ffbe090217b806221c8053cc5b521bd45bc7047eaccdfc5ef96"} Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.133005 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerStarted","Data":"7345a5a95dcc990570fff46fd44e56410603324ddbd092ceb1895f9ddb4c326f"} Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.278044 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.284254 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8cce4f5d-78aa-44c4-8c60-8bce681ada4e-webhook-certs\") pod \"openstack-operator-controller-manager-6cf56d77c9-f7stz\" (UID: \"8cce4f5d-78aa-44c4-8c60-8bce681ada4e\") " pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.480682 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:42 crc kubenswrapper[4972]: I0228 10:51:42.708099 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz"] Feb 28 10:51:42 crc kubenswrapper[4972]: W0228 10:51:42.715520 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cce4f5d_78aa_44c4_8c60_8bce681ada4e.slice/crio-e19c16cecfb035f7820e559ee04edfddaa503ca0aa8147ae7b96421270d560b1 WatchSource:0}: Error finding container e19c16cecfb035f7820e559ee04edfddaa503ca0aa8147ae7b96421270d560b1: Status 404 returned error can't find the container with id e19c16cecfb035f7820e559ee04edfddaa503ca0aa8147ae7b96421270d560b1 Feb 28 10:51:43 crc kubenswrapper[4972]: I0228 10:51:43.143039 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" event={"ID":"8cce4f5d-78aa-44c4-8c60-8bce681ada4e","Type":"ContainerStarted","Data":"e19c16cecfb035f7820e559ee04edfddaa503ca0aa8147ae7b96421270d560b1"} Feb 28 10:51:43 crc kubenswrapper[4972]: I0228 10:51:43.145709 4972 generic.go:334] "Generic (PLEG): container finished" podID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerID="becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b" exitCode=0 Feb 28 10:51:43 crc kubenswrapper[4972]: I0228 10:51:43.145779 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerDied","Data":"becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b"} Feb 28 10:51:44 crc kubenswrapper[4972]: I0228 10:51:44.156957 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" event={"ID":"8cce4f5d-78aa-44c4-8c60-8bce681ada4e","Type":"ContainerStarted","Data":"ff81b9a03b943b755fe0f0b95d0d15b5885d0a78aa0e50e255af939d762f0148"} Feb 28 10:51:45 crc kubenswrapper[4972]: I0228 10:51:45.164355 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:45 crc kubenswrapper[4972]: I0228 10:51:45.195926 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" podStartSLOduration=35.195891501 podStartE2EDuration="35.195891501s" podCreationTimestamp="2026-02-28 10:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:51:45.194108751 +0000 UTC m=+962.106116499" watchObservedRunningTime="2026-02-28 10:51:45.195891501 +0000 UTC m=+962.107899239" Feb 28 10:51:46 crc kubenswrapper[4972]: I0228 10:51:46.288827 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw" Feb 28 10:51:47 crc kubenswrapper[4972]: I0228 10:51:47.189374 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" event={"ID":"15d3191f-413a-4806-8e55-9dd19db4c335","Type":"ContainerStarted","Data":"34368750b156e6b79bb81cc0001798d4121f8f2bad2f4eff480d54c4f2118221"} Feb 28 10:51:47 crc kubenswrapper[4972]: I0228 10:51:47.191091 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:51:47 crc kubenswrapper[4972]: I0228 10:51:47.193110 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerStarted","Data":"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552"} Feb 28 10:51:47 crc kubenswrapper[4972]: I0228 10:51:47.214864 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" podStartSLOduration=3.332484801 podStartE2EDuration="38.214833867s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:11.558529371 +0000 UTC m=+928.470537099" lastFinishedPulling="2026-02-28 10:51:46.440878427 +0000 UTC m=+963.352886165" observedRunningTime="2026-02-28 10:51:47.209964861 +0000 UTC m=+964.121972639" watchObservedRunningTime="2026-02-28 10:51:47.214833867 +0000 UTC m=+964.126841605" Feb 28 10:51:48 crc kubenswrapper[4972]: I0228 10:51:48.211221 4972 generic.go:334] "Generic (PLEG): container finished" podID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerID="0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552" exitCode=0 Feb 28 10:51:48 crc kubenswrapper[4972]: I0228 10:51:48.211314 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerDied","Data":"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552"} Feb 28 10:51:49 crc kubenswrapper[4972]: I0228 10:51:49.219646 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" event={"ID":"0c55bc45-0607-477d-893f-5782b00cffb2","Type":"ContainerStarted","Data":"64c406f52cd5c6e1db08041c2ae31342820f5dcd65a171daa8001b0c8f52dd82"} Feb 28 10:51:49 crc kubenswrapper[4972]: I0228 10:51:49.220282 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:51:49 crc kubenswrapper[4972]: I0228 10:51:49.225387 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerStarted","Data":"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef"} Feb 28 10:51:49 crc kubenswrapper[4972]: I0228 10:51:49.243007 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" podStartSLOduration=33.897812747 podStartE2EDuration="40.242960748s" podCreationTimestamp="2026-02-28 10:51:09 +0000 UTC" firstStartedPulling="2026-02-28 10:51:42.096298631 +0000 UTC m=+959.008306369" lastFinishedPulling="2026-02-28 10:51:48.441446632 +0000 UTC m=+965.353454370" observedRunningTime="2026-02-28 10:51:49.239358257 +0000 UTC m=+966.151365995" watchObservedRunningTime="2026-02-28 10:51:49.242960748 +0000 UTC m=+966.154968486" Feb 28 10:51:49 crc kubenswrapper[4972]: I0228 10:51:49.265490 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-24lg8" podStartSLOduration=3.6399354 podStartE2EDuration="9.265447173s" podCreationTimestamp="2026-02-28 10:51:40 +0000 UTC" firstStartedPulling="2026-02-28 10:51:43.147908627 +0000 UTC m=+960.059916365" lastFinishedPulling="2026-02-28 10:51:48.7734204 +0000 UTC m=+965.685428138" observedRunningTime="2026-02-28 10:51:49.259764885 +0000 UTC m=+966.171772633" watchObservedRunningTime="2026-02-28 10:51:49.265447173 +0000 UTC m=+966.177454911" Feb 28 10:51:50 crc kubenswrapper[4972]: I0228 10:51:50.934769 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:50 crc kubenswrapper[4972]: I0228 10:51:50.934843 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:51.999441 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-24lg8" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="registry-server" probeResult="failure" output=< Feb 28 10:51:52 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:51:52 crc kubenswrapper[4972]: > Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.490109 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cf56d77c9-f7stz" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.641163 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.642680 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.675497 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.763138 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pc8j\" (UniqueName: \"kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.763206 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.763233 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.864667 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.864730 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.864836 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pc8j\" (UniqueName: \"kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.865766 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.866007 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.888199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pc8j\" (UniqueName: \"kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j\") pod \"redhat-marketplace-42lq6\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:52 crc kubenswrapper[4972]: I0228 10:51:52.964958 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:51:53 crc kubenswrapper[4972]: I0228 10:51:53.465097 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:51:53 crc kubenswrapper[4972]: W0228 10:51:53.469290 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61d10a8b_9a31_4d9f_b59c_e1573f904f1b.slice/crio-c12fd71eba0e8e88fff942e849881cf36ace2a69d0dec2ce3eb7e5b56f3def59 WatchSource:0}: Error finding container c12fd71eba0e8e88fff942e849881cf36ace2a69d0dec2ce3eb7e5b56f3def59: Status 404 returned error can't find the container with id c12fd71eba0e8e88fff942e849881cf36ace2a69d0dec2ce3eb7e5b56f3def59 Feb 28 10:51:54 crc kubenswrapper[4972]: I0228 10:51:54.266068 4972 generic.go:334] "Generic (PLEG): container finished" podID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerID="066a6a665cb4a0bdd35e819d6b2152b3f51872fc508bab54a1630f33deac00e6" exitCode=0 Feb 28 10:51:54 crc kubenswrapper[4972]: I0228 10:51:54.266149 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerDied","Data":"066a6a665cb4a0bdd35e819d6b2152b3f51872fc508bab54a1630f33deac00e6"} Feb 28 10:51:54 crc kubenswrapper[4972]: I0228 10:51:54.266612 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerStarted","Data":"c12fd71eba0e8e88fff942e849881cf36ace2a69d0dec2ce3eb7e5b56f3def59"} Feb 28 10:51:55 crc kubenswrapper[4972]: I0228 10:51:55.278098 4972 generic.go:334] "Generic (PLEG): container finished" podID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerID="8c41b5bf62eb31af5f6929a45b1b860e3cd0ee5e8611e9543ae814e64f804eca" exitCode=0 Feb 28 10:51:55 crc kubenswrapper[4972]: I0228 10:51:55.278222 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerDied","Data":"8c41b5bf62eb31af5f6929a45b1b860e3cd0ee5e8611e9543ae814e64f804eca"} Feb 28 10:51:56 crc kubenswrapper[4972]: I0228 10:51:56.289789 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerStarted","Data":"295f38ef45d4a6c65e67ebf424360cc08c902ce099d28e72b4c6dfba6860de54"} Feb 28 10:51:59 crc kubenswrapper[4972]: I0228 10:51:59.982340 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-78b64779b9-xt2pl" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.006109 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-42lq6" podStartSLOduration=6.353615589 podStartE2EDuration="8.006076806s" podCreationTimestamp="2026-02-28 10:51:52 +0000 UTC" firstStartedPulling="2026-02-28 10:51:54.26872259 +0000 UTC m=+971.180730328" lastFinishedPulling="2026-02-28 10:51:55.921183817 +0000 UTC m=+972.833191545" observedRunningTime="2026-02-28 10:51:56.331201777 +0000 UTC m=+973.243209515" watchObservedRunningTime="2026-02-28 10:52:00.006076806 +0000 UTC m=+976.918084564" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.143286 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537932-5ql5s"] Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.144218 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.146748 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.146988 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.147152 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.154305 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537932-5ql5s"] Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.186687 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwm6b\" (UniqueName: \"kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b\") pod \"auto-csr-approver-29537932-5ql5s\" (UID: \"58a7b699-f95c-43cd-b919-8ae540010828\") " pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.287817 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwm6b\" (UniqueName: \"kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b\") pod \"auto-csr-approver-29537932-5ql5s\" (UID: \"58a7b699-f95c-43cd-b919-8ae540010828\") " pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.310342 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwm6b\" (UniqueName: \"kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b\") pod \"auto-csr-approver-29537932-5ql5s\" (UID: \"58a7b699-f95c-43cd-b919-8ae540010828\") " pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.469008 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.744104 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537932-5ql5s"] Feb 28 10:52:00 crc kubenswrapper[4972]: W0228 10:52:00.750694 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58a7b699_f95c_43cd_b919_8ae540010828.slice/crio-3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f WatchSource:0}: Error finding container 3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f: Status 404 returned error can't find the container with id 3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f Feb 28 10:52:00 crc kubenswrapper[4972]: I0228 10:52:00.985511 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:52:01 crc kubenswrapper[4972]: I0228 10:52:01.036679 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:52:01 crc kubenswrapper[4972]: I0228 10:52:01.232975 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:52:01 crc kubenswrapper[4972]: I0228 10:52:01.335670 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" event={"ID":"58a7b699-f95c-43cd-b919-8ae540010828","Type":"ContainerStarted","Data":"3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f"} Feb 28 10:52:01 crc kubenswrapper[4972]: I0228 10:52:01.570211 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-c77466965-bkt68" Feb 28 10:52:02 crc kubenswrapper[4972]: I0228 10:52:02.342148 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-24lg8" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="registry-server" containerID="cri-o://ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef" gracePeriod=2 Feb 28 10:52:02 crc kubenswrapper[4972]: I0228 10:52:02.965431 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:02 crc kubenswrapper[4972]: I0228 10:52:02.966036 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.019192 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.317167 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.360427 4972 generic.go:334] "Generic (PLEG): container finished" podID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerID="ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef" exitCode=0 Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.360512 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24lg8" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.360515 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerDied","Data":"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef"} Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.360608 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24lg8" event={"ID":"ef2c576e-443c-462e-bbe3-ce42107bae83","Type":"ContainerDied","Data":"7345a5a95dcc990570fff46fd44e56410603324ddbd092ceb1895f9ddb4c326f"} Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.360632 4972 scope.go:117] "RemoveContainer" containerID="ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.363818 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" event={"ID":"58a7b699-f95c-43cd-b919-8ae540010828","Type":"ContainerStarted","Data":"e03ec78b341831c5f5f29feb429aef7b566d4988575ed872c29341e63e13117e"} Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.416551 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.425388 4972 scope.go:117] "RemoveContainer" containerID="0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.440558 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities\") pod \"ef2c576e-443c-462e-bbe3-ce42107bae83\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.440634 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d484f\" (UniqueName: \"kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f\") pod \"ef2c576e-443c-462e-bbe3-ce42107bae83\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.440845 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content\") pod \"ef2c576e-443c-462e-bbe3-ce42107bae83\" (UID: \"ef2c576e-443c-462e-bbe3-ce42107bae83\") " Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.443041 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities" (OuterVolumeSpecName: "utilities") pod "ef2c576e-443c-462e-bbe3-ce42107bae83" (UID: "ef2c576e-443c-462e-bbe3-ce42107bae83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.449379 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f" (OuterVolumeSpecName: "kube-api-access-d484f") pod "ef2c576e-443c-462e-bbe3-ce42107bae83" (UID: "ef2c576e-443c-462e-bbe3-ce42107bae83"). InnerVolumeSpecName "kube-api-access-d484f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.456240 4972 scope.go:117] "RemoveContainer" containerID="becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.491277 4972 scope.go:117] "RemoveContainer" containerID="ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef" Feb 28 10:52:03 crc kubenswrapper[4972]: E0228 10:52:03.492214 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef\": container with ID starting with ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef not found: ID does not exist" containerID="ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.492271 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef"} err="failed to get container status \"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef\": rpc error: code = NotFound desc = could not find container \"ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef\": container with ID starting with ad4506e47d9c9406fb220723584272fe3833348b94b5a73265a38f365c22c2ef not found: ID does not exist" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.492310 4972 scope.go:117] "RemoveContainer" containerID="0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552" Feb 28 10:52:03 crc kubenswrapper[4972]: E0228 10:52:03.492937 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552\": container with ID starting with 0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552 not found: ID does not exist" containerID="0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.492966 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552"} err="failed to get container status \"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552\": rpc error: code = NotFound desc = could not find container \"0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552\": container with ID starting with 0e573268eb84336a9f7a7bfe2f7947b9ca688ab552df68c18d621b4ca5cec552 not found: ID does not exist" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.492997 4972 scope.go:117] "RemoveContainer" containerID="becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b" Feb 28 10:52:03 crc kubenswrapper[4972]: E0228 10:52:03.493670 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b\": container with ID starting with becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b not found: ID does not exist" containerID="becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.493723 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b"} err="failed to get container status \"becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b\": rpc error: code = NotFound desc = could not find container \"becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b\": container with ID starting with becc10e86e0645f9dccfe715dadf41be7b73ac91b4d7e8a6ac58c72b5d30220b not found: ID does not exist" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.506771 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef2c576e-443c-462e-bbe3-ce42107bae83" (UID: "ef2c576e-443c-462e-bbe3-ce42107bae83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.542951 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.542992 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2c576e-443c-462e-bbe3-ce42107bae83-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.543004 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d484f\" (UniqueName: \"kubernetes.io/projected/ef2c576e-443c-462e-bbe3-ce42107bae83-kube-api-access-d484f\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.709599 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.719175 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-24lg8"] Feb 28 10:52:03 crc kubenswrapper[4972]: I0228 10:52:03.804359 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" path="/var/lib/kubelet/pods/ef2c576e-443c-462e-bbe3-ce42107bae83/volumes" Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.391429 4972 generic.go:334] "Generic (PLEG): container finished" podID="58a7b699-f95c-43cd-b919-8ae540010828" containerID="e03ec78b341831c5f5f29feb429aef7b566d4988575ed872c29341e63e13117e" exitCode=0 Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.391539 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" event={"ID":"58a7b699-f95c-43cd-b919-8ae540010828","Type":"ContainerDied","Data":"e03ec78b341831c5f5f29feb429aef7b566d4988575ed872c29341e63e13117e"} Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.665645 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.761355 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwm6b\" (UniqueName: \"kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b\") pod \"58a7b699-f95c-43cd-b919-8ae540010828\" (UID: \"58a7b699-f95c-43cd-b919-8ae540010828\") " Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.766072 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b" (OuterVolumeSpecName: "kube-api-access-gwm6b") pod "58a7b699-f95c-43cd-b919-8ae540010828" (UID: "58a7b699-f95c-43cd-b919-8ae540010828"). InnerVolumeSpecName "kube-api-access-gwm6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:04 crc kubenswrapper[4972]: I0228 10:52:04.862749 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwm6b\" (UniqueName: \"kubernetes.io/projected/58a7b699-f95c-43cd-b919-8ae540010828-kube-api-access-gwm6b\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.404840 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" event={"ID":"58a7b699-f95c-43cd-b919-8ae540010828","Type":"ContainerDied","Data":"3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f"} Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.405186 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f2f0ea430866d47bfe37a1882cfcc71ed61b2a936138116cbc9a4f519a1b98f" Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.404905 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537932-5ql5s" Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.743169 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537926-krdlc"] Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.748233 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537926-krdlc"] Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.799772 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="953c17df-68d7-407d-b186-17551b7b97b2" path="/var/lib/kubelet/pods/953c17df-68d7-407d-b186-17551b7b97b2/volumes" Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.838968 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:52:05 crc kubenswrapper[4972]: I0228 10:52:05.839210 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-42lq6" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="registry-server" containerID="cri-o://295f38ef45d4a6c65e67ebf424360cc08c902ce099d28e72b4c6dfba6860de54" gracePeriod=2 Feb 28 10:52:07 crc kubenswrapper[4972]: I0228 10:52:07.424704 4972 generic.go:334] "Generic (PLEG): container finished" podID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerID="295f38ef45d4a6c65e67ebf424360cc08c902ce099d28e72b4c6dfba6860de54" exitCode=0 Feb 28 10:52:07 crc kubenswrapper[4972]: I0228 10:52:07.424779 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerDied","Data":"295f38ef45d4a6c65e67ebf424360cc08c902ce099d28e72b4c6dfba6860de54"} Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.428933 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.435435 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-42lq6" event={"ID":"61d10a8b-9a31-4d9f-b59c-e1573f904f1b","Type":"ContainerDied","Data":"c12fd71eba0e8e88fff942e849881cf36ace2a69d0dec2ce3eb7e5b56f3def59"} Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.435516 4972 scope.go:117] "RemoveContainer" containerID="295f38ef45d4a6c65e67ebf424360cc08c902ce099d28e72b4c6dfba6860de54" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.435540 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-42lq6" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.541901 4972 scope.go:117] "RemoveContainer" containerID="8c41b5bf62eb31af5f6929a45b1b860e3cd0ee5e8611e9543ae814e64f804eca" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.566340 4972 scope.go:117] "RemoveContainer" containerID="066a6a665cb4a0bdd35e819d6b2152b3f51872fc508bab54a1630f33deac00e6" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.620337 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities\") pod \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.620426 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pc8j\" (UniqueName: \"kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j\") pod \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.620505 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content\") pod \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\" (UID: \"61d10a8b-9a31-4d9f-b59c-e1573f904f1b\") " Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.621704 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities" (OuterVolumeSpecName: "utilities") pod "61d10a8b-9a31-4d9f-b59c-e1573f904f1b" (UID: "61d10a8b-9a31-4d9f-b59c-e1573f904f1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.627684 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j" (OuterVolumeSpecName: "kube-api-access-4pc8j") pod "61d10a8b-9a31-4d9f-b59c-e1573f904f1b" (UID: "61d10a8b-9a31-4d9f-b59c-e1573f904f1b"). InnerVolumeSpecName "kube-api-access-4pc8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.646921 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61d10a8b-9a31-4d9f-b59c-e1573f904f1b" (UID: "61d10a8b-9a31-4d9f-b59c-e1573f904f1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.722208 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pc8j\" (UniqueName: \"kubernetes.io/projected/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-kube-api-access-4pc8j\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.722240 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.722251 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61d10a8b-9a31-4d9f-b59c-e1573f904f1b-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.780617 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:52:08 crc kubenswrapper[4972]: I0228 10:52:08.791003 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-42lq6"] Feb 28 10:52:09 crc kubenswrapper[4972]: I0228 10:52:09.798299 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" path="/var/lib/kubelet/pods/61d10a8b-9a31-4d9f-b59c-e1573f904f1b/volumes" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.297597 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298656 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a7b699-f95c-43cd-b919-8ae540010828" containerName="oc" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298678 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a7b699-f95c-43cd-b919-8ae540010828" containerName="oc" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298698 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="extract-content" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298707 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="extract-content" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298724 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298734 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298746 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298754 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298768 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="extract-utilities" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298777 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="extract-utilities" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298794 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="extract-utilities" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298802 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="extract-utilities" Feb 28 10:52:19 crc kubenswrapper[4972]: E0228 10:52:19.298812 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="extract-content" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.298821 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="extract-content" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.299043 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d10a8b-9a31-4d9f-b59c-e1573f904f1b" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.299063 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2c576e-443c-462e-bbe3-ce42107bae83" containerName="registry-server" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.299083 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a7b699-f95c-43cd-b919-8ae540010828" containerName="oc" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.301329 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.304137 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rpmmc" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.305513 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.305634 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.309098 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.323742 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.409223 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.410529 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.413424 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.426773 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.502567 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhhmk\" (UniqueName: \"kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.502647 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.604276 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhhmk\" (UniqueName: \"kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.604371 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qf7\" (UniqueName: \"kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.604424 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.604554 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.604716 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.606641 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.631320 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhhmk\" (UniqueName: \"kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk\") pod \"dnsmasq-dns-675f4bcbfc-nwt8p\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.705711 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.705818 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qf7\" (UniqueName: \"kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.705845 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.706987 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.707020 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.723239 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qf7\" (UniqueName: \"kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7\") pod \"dnsmasq-dns-78dd6ddcc-bx4mj\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.727134 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:19 crc kubenswrapper[4972]: I0228 10:52:19.921711 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:20 crc kubenswrapper[4972]: I0228 10:52:20.166066 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:20 crc kubenswrapper[4972]: W0228 10:52:20.168795 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66bcf78c_0ac0_453e_9a95_f55320054ad9.slice/crio-58faa5fe5df7d51a24637ced099efcf45e62284807890d35bc1f1bb1f77da612 WatchSource:0}: Error finding container 58faa5fe5df7d51a24637ced099efcf45e62284807890d35bc1f1bb1f77da612: Status 404 returned error can't find the container with id 58faa5fe5df7d51a24637ced099efcf45e62284807890d35bc1f1bb1f77da612 Feb 28 10:52:20 crc kubenswrapper[4972]: I0228 10:52:20.203683 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:20 crc kubenswrapper[4972]: I0228 10:52:20.532207 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" event={"ID":"9a71e95a-5289-47ba-aaf9-5574c65d4c7b","Type":"ContainerStarted","Data":"f143360d6c5b9d6595c9daaa37318f374c2cb2f135137ca6fc7033dc2ae19d8f"} Feb 28 10:52:20 crc kubenswrapper[4972]: I0228 10:52:20.534443 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" event={"ID":"66bcf78c-0ac0-453e-9a95-f55320054ad9","Type":"ContainerStarted","Data":"58faa5fe5df7d51a24637ced099efcf45e62284807890d35bc1f1bb1f77da612"} Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.050813 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.075446 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.077098 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.089765 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.249917 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.249989 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wzks\" (UniqueName: \"kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.250023 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.353372 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wzks\" (UniqueName: \"kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.353488 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.353584 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.354981 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.362394 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.388403 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wzks\" (UniqueName: \"kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks\") pod \"dnsmasq-dns-5ccc8479f9-f6vp4\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.389168 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.406290 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.408505 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.409813 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.423145 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.557279 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wngkw\" (UniqueName: \"kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.557799 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.557937 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.664473 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.664993 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.665037 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wngkw\" (UniqueName: \"kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.666598 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.667154 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.694357 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wngkw\" (UniqueName: \"kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw\") pod \"dnsmasq-dns-57d769cc4f-wv627\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.756982 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:22 crc kubenswrapper[4972]: I0228 10:52:22.979294 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:22 crc kubenswrapper[4972]: W0228 10:52:22.992440 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08476c57_ac49_4644_a9e6_e81340e19370.slice/crio-f9672a0a03b139bcf6e6d174e0cfd6a8701e9f0be492356f4717c4b763da7d02 WatchSource:0}: Error finding container f9672a0a03b139bcf6e6d174e0cfd6a8701e9f0be492356f4717c4b763da7d02: Status 404 returned error can't find the container with id f9672a0a03b139bcf6e6d174e0cfd6a8701e9f0be492356f4717c4b763da7d02 Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.237426 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:52:23 crc kubenswrapper[4972]: W0228 10:52:23.240826 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda43ffba7_dbee_4012_b991_4c809386fcf6.slice/crio-e274afcd0c606a52efcd1ab0331a8c8de1ccae4ecbfef0b8e52e40a8a6c3a9e5 WatchSource:0}: Error finding container e274afcd0c606a52efcd1ab0331a8c8de1ccae4ecbfef0b8e52e40a8a6c3a9e5: Status 404 returned error can't find the container with id e274afcd0c606a52efcd1ab0331a8c8de1ccae4ecbfef0b8e52e40a8a6c3a9e5 Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.247964 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.249928 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.254618 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.254692 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.255716 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.256001 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.256166 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zgbf4" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.256993 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.257105 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.266125 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377028 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377078 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377100 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377119 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxmdw\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377135 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377156 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377197 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377224 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377384 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377441 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.377492 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479024 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479098 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479124 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479150 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479176 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479200 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479220 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxmdw\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479249 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479296 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479330 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479367 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.479870 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.484350 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.484375 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.484745 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.485175 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.485398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.490449 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.490625 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.491179 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.498165 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.500832 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxmdw\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.504041 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.551107 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.556823 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563167 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563273 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jxt7b" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563183 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563365 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563421 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563453 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.563638 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.565064 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.588976 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.599846 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" event={"ID":"a43ffba7-dbee-4012-b991-4c809386fcf6","Type":"ContainerStarted","Data":"e274afcd0c606a52efcd1ab0331a8c8de1ccae4ecbfef0b8e52e40a8a6c3a9e5"} Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.602421 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" event={"ID":"08476c57-ac49-4644-a9e6-e81340e19370","Type":"ContainerStarted","Data":"f9672a0a03b139bcf6e6d174e0cfd6a8701e9f0be492356f4717c4b763da7d02"} Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693176 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693428 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693480 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693513 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693536 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2w9p\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693816 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.693884 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.694046 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.694076 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.694217 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.694278 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795592 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795663 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795687 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795710 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795777 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2w9p\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795794 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795820 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795855 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795870 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.795895 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.796381 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.796679 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.797124 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.797326 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.797565 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.799395 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.803106 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.813420 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.822797 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.829614 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2w9p\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.835774 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.840744 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " pod="openstack/rabbitmq-server-0" Feb 28 10:52:23 crc kubenswrapper[4972]: I0228 10:52:23.913394 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.887521 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.889567 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.893219 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.895099 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8th7w" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.895175 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.895409 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.900147 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 28 10:52:24 crc kubenswrapper[4972]: I0228 10:52:24.921479 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.018436 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-operator-scripts\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.018607 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-kolla-config\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.018651 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.018677 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mtkz\" (UniqueName: \"kubernetes.io/projected/57d374b4-97c1-4b80-8f4b-95338fe4f385-kube-api-access-5mtkz\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.018822 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-default\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.019017 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.019179 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-generated\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.019228 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121312 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121395 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-generated\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121415 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-operator-scripts\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121514 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-kolla-config\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121548 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121572 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mtkz\" (UniqueName: \"kubernetes.io/projected/57d374b4-97c1-4b80-8f4b-95338fe4f385-kube-api-access-5mtkz\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.121609 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-default\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.122653 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-default\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.123974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57d374b4-97c1-4b80-8f4b-95338fe4f385-config-data-generated\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.124650 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-operator-scripts\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.124798 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.127170 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.127863 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57d374b4-97c1-4b80-8f4b-95338fe4f385-kolla-config\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.127916 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57d374b4-97c1-4b80-8f4b-95338fe4f385-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.150957 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mtkz\" (UniqueName: \"kubernetes.io/projected/57d374b4-97c1-4b80-8f4b-95338fe4f385-kube-api-access-5mtkz\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.156871 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"57d374b4-97c1-4b80-8f4b-95338fe4f385\") " pod="openstack/openstack-galera-0" Feb 28 10:52:25 crc kubenswrapper[4972]: I0228 10:52:25.213018 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.308860 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.310116 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.315091 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.316563 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.316940 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xsptk" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.317115 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.353762 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l6tx\" (UniqueName: \"kubernetes.io/projected/44e7c813-beaa-4015-a581-afcaea85077c-kube-api-access-7l6tx\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.354375 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44e7c813-beaa-4015-a581-afcaea85077c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.354430 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.354522 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.354945 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.355045 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.355146 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.355590 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.361025 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457007 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457113 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457162 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l6tx\" (UniqueName: \"kubernetes.io/projected/44e7c813-beaa-4015-a581-afcaea85077c-kube-api-access-7l6tx\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457211 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44e7c813-beaa-4015-a581-afcaea85077c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457277 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457311 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.457335 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.458137 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.458262 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.459360 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44e7c813-beaa-4015-a581-afcaea85077c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.460838 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.460962 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e7c813-beaa-4015-a581-afcaea85077c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.466878 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.466876 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7c813-beaa-4015-a581-afcaea85077c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.484873 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.488600 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l6tx\" (UniqueName: \"kubernetes.io/projected/44e7c813-beaa-4015-a581-afcaea85077c-kube-api-access-7l6tx\") pod \"openstack-cell1-galera-0\" (UID: \"44e7c813-beaa-4015-a581-afcaea85077c\") " pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.548738 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.556250 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.560229 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qp9qf" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.560634 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.560849 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.569348 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.651785 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.660290 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-config-data\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.660366 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.660391 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-kolla-config\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.660418 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.660445 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlrfj\" (UniqueName: \"kubernetes.io/projected/56e07ece-d896-4107-b9b1-6f24b64f7985-kube-api-access-qlrfj\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.761985 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-config-data\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.762070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.762096 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-kolla-config\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.762121 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.762147 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlrfj\" (UniqueName: \"kubernetes.io/projected/56e07ece-d896-4107-b9b1-6f24b64f7985-kube-api-access-qlrfj\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.763418 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-config-data\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.763684 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56e07ece-d896-4107-b9b1-6f24b64f7985-kolla-config\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.766378 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-combined-ca-bundle\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.766569 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e07ece-d896-4107-b9b1-6f24b64f7985-memcached-tls-certs\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.790170 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlrfj\" (UniqueName: \"kubernetes.io/projected/56e07ece-d896-4107-b9b1-6f24b64f7985-kube-api-access-qlrfj\") pod \"memcached-0\" (UID: \"56e07ece-d896-4107-b9b1-6f24b64f7985\") " pod="openstack/memcached-0" Feb 28 10:52:26 crc kubenswrapper[4972]: I0228 10:52:26.874641 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.113481 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.119164 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.128440 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.131364 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-jrtjm" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.210142 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scdw9\" (UniqueName: \"kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9\") pod \"kube-state-metrics-0\" (UID: \"daacfde0-2575-4448-a11f-0eb9a1b2a1cb\") " pod="openstack/kube-state-metrics-0" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.311858 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scdw9\" (UniqueName: \"kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9\") pod \"kube-state-metrics-0\" (UID: \"daacfde0-2575-4448-a11f-0eb9a1b2a1cb\") " pod="openstack/kube-state-metrics-0" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.331948 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scdw9\" (UniqueName: \"kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9\") pod \"kube-state-metrics-0\" (UID: \"daacfde0-2575-4448-a11f-0eb9a1b2a1cb\") " pod="openstack/kube-state-metrics-0" Feb 28 10:52:29 crc kubenswrapper[4972]: I0228 10:52:29.448344 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.334968 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.336848 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.342673 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.343107 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.343161 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.346005 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.346097 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-pmp79" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.356900 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377685 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377788 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-config\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377895 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377937 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.377965 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.378177 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.378254 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxmds\" (UniqueName: \"kubernetes.io/projected/6ba1065c-4ef3-473c-9576-d673f19a75a1-kube-api-access-vxmds\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.479984 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.480046 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxmds\" (UniqueName: \"kubernetes.io/projected/6ba1065c-4ef3-473c-9576-d673f19a75a1-kube-api-access-vxmds\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.480613 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.480652 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.480696 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-config\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.481246 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.481299 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.481328 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.481997 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-config\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.482281 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6ba1065c-4ef3-473c-9576-d673f19a75a1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.482777 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.483279 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.489269 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.489321 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.498381 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ba1065c-4ef3-473c-9576-d673f19a75a1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.503948 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxmds\" (UniqueName: \"kubernetes.io/projected/6ba1065c-4ef3-473c-9576-d673f19a75a1-kube-api-access-vxmds\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.506286 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6ba1065c-4ef3-473c-9576-d673f19a75a1\") " pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:32 crc kubenswrapper[4972]: I0228 10:52:32.665901 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.269436 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p6jmv"] Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.271391 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.274977 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.275040 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hfvsh" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.274978 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.286272 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p6jmv"] Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308609 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf6nb\" (UniqueName: \"kubernetes.io/projected/d90126a7-5923-45bd-9e61-29d1e2e3f469-kube-api-access-lf6nb\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308677 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308704 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-combined-ca-bundle\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308731 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-ovn-controller-tls-certs\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308785 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308805 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d90126a7-5923-45bd-9e61-29d1e2e3f469-scripts\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.308853 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-log-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.360293 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5wgfj"] Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.361964 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.378630 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5wgfj"] Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.410842 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.410994 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d90126a7-5923-45bd-9e61-29d1e2e3f469-scripts\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411050 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-log-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411079 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-log\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411101 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9l9w\" (UniqueName: \"kubernetes.io/projected/7eb839e4-8819-45ba-9afd-312cc02a6964-kube-api-access-h9l9w\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411129 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-run\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411163 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf6nb\" (UniqueName: \"kubernetes.io/projected/d90126a7-5923-45bd-9e61-29d1e2e3f469-kube-api-access-lf6nb\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411197 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eb839e4-8819-45ba-9afd-312cc02a6964-scripts\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411218 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411238 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-combined-ca-bundle\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411262 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-ovn-controller-tls-certs\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411300 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-etc-ovs\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.411322 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-lib\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.414789 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d90126a7-5923-45bd-9e61-29d1e2e3f469-scripts\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.415287 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-log-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.415393 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.415539 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d90126a7-5923-45bd-9e61-29d1e2e3f469-var-run-ovn\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.423669 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-combined-ca-bundle\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.428025 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d90126a7-5923-45bd-9e61-29d1e2e3f469-ovn-controller-tls-certs\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.435703 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf6nb\" (UniqueName: \"kubernetes.io/projected/d90126a7-5923-45bd-9e61-29d1e2e3f469-kube-api-access-lf6nb\") pod \"ovn-controller-p6jmv\" (UID: \"d90126a7-5923-45bd-9e61-29d1e2e3f469\") " pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513093 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-log\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513167 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9l9w\" (UniqueName: \"kubernetes.io/projected/7eb839e4-8819-45ba-9afd-312cc02a6964-kube-api-access-h9l9w\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513221 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-run\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513291 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eb839e4-8819-45ba-9afd-312cc02a6964-scripts\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513355 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-etc-ovs\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513384 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-lib\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513424 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-log\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513535 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-run\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513643 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-etc-ovs\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.513710 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7eb839e4-8819-45ba-9afd-312cc02a6964-var-lib\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.516268 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eb839e4-8819-45ba-9afd-312cc02a6964-scripts\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.536058 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9l9w\" (UniqueName: \"kubernetes.io/projected/7eb839e4-8819-45ba-9afd-312cc02a6964-kube-api-access-h9l9w\") pod \"ovn-controller-ovs-5wgfj\" (UID: \"7eb839e4-8819-45ba-9afd-312cc02a6964\") " pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.608680 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:33 crc kubenswrapper[4972]: I0228 10:52:33.694683 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.918716 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.920642 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.924889 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.925220 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-nz9xb" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.925800 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.926654 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.934301 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963116 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963178 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963241 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963264 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963312 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963340 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kblh\" (UniqueName: \"kubernetes.io/projected/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-kube-api-access-6kblh\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963370 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:35 crc kubenswrapper[4972]: I0228 10:52:35.963392 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.073836 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.073967 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.073998 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074049 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074081 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kblh\" (UniqueName: \"kubernetes.io/projected/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-kube-api-access-6kblh\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074109 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074132 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074215 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.074710 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.075703 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.080366 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.080667 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-config\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.084014 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.087398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.095305 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.100132 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kblh\" (UniqueName: \"kubernetes.io/projected/ec76b1b5-ea80-49cb-ae7c-0190a52640a1-kube-api-access-6kblh\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.100482 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ec76b1b5-ea80-49cb-ae7c-0190a52640a1\") " pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: I0228 10:52:36.273912 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.602606 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.602794 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7wzks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-f6vp4_openstack(08476c57-ac49-4644-a9e6-e81340e19370): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.603992 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" podUID="08476c57-ac49-4644-a9e6-e81340e19370" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.613677 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.613845 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9qf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-bx4mj_openstack(66bcf78c-0ac0-453e-9a95-f55320054ad9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.615075 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" podUID="66bcf78c-0ac0-453e-9a95-f55320054ad9" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.666335 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.666602 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fhhmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-nwt8p_openstack(9a71e95a-5289-47ba-aaf9-5574c65d4c7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:52:36 crc kubenswrapper[4972]: E0228 10:52:36.667743 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" podUID="9a71e95a-5289-47ba-aaf9-5574c65d4c7b" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.353754 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.355293 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.460591 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.473981 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: W0228 10:52:37.489787 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffd40994_f4d6_442c_97fa_6ed417500617.slice/crio-aa7446a5a43998a2fbee8e2ee785ebdb631ce8107b26032d4d3066cccc55bdf0 WatchSource:0}: Error finding container aa7446a5a43998a2fbee8e2ee785ebdb631ce8107b26032d4d3066cccc55bdf0: Status 404 returned error can't find the container with id aa7446a5a43998a2fbee8e2ee785ebdb631ce8107b26032d4d3066cccc55bdf0 Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.490816 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.503240 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.509783 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514131 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config\") pod \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514205 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc\") pod \"66bcf78c-0ac0-453e-9a95-f55320054ad9\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514248 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9qf7\" (UniqueName: \"kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7\") pod \"66bcf78c-0ac0-453e-9a95-f55320054ad9\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514278 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhhmk\" (UniqueName: \"kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk\") pod \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\" (UID: \"9a71e95a-5289-47ba-aaf9-5574c65d4c7b\") " Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514357 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config\") pod \"66bcf78c-0ac0-453e-9a95-f55320054ad9\" (UID: \"66bcf78c-0ac0-453e-9a95-f55320054ad9\") " Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514667 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config" (OuterVolumeSpecName: "config") pod "9a71e95a-5289-47ba-aaf9-5574c65d4c7b" (UID: "9a71e95a-5289-47ba-aaf9-5574c65d4c7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.514765 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.515181 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config" (OuterVolumeSpecName: "config") pod "66bcf78c-0ac0-453e-9a95-f55320054ad9" (UID: "66bcf78c-0ac0-453e-9a95-f55320054ad9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.515979 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66bcf78c-0ac0-453e-9a95-f55320054ad9" (UID: "66bcf78c-0ac0-453e-9a95-f55320054ad9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.520664 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk" (OuterVolumeSpecName: "kube-api-access-fhhmk") pod "9a71e95a-5289-47ba-aaf9-5574c65d4c7b" (UID: "9a71e95a-5289-47ba-aaf9-5574c65d4c7b"). InnerVolumeSpecName "kube-api-access-fhhmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.521061 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7" (OuterVolumeSpecName: "kube-api-access-l9qf7") pod "66bcf78c-0ac0-453e-9a95-f55320054ad9" (UID: "66bcf78c-0ac0-453e-9a95-f55320054ad9"). InnerVolumeSpecName "kube-api-access-l9qf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.617478 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.617552 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9qf7\" (UniqueName: \"kubernetes.io/projected/66bcf78c-0ac0-453e-9a95-f55320054ad9-kube-api-access-l9qf7\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.617573 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhhmk\" (UniqueName: \"kubernetes.io/projected/9a71e95a-5289-47ba-aaf9-5574c65d4c7b-kube-api-access-fhhmk\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.617583 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66bcf78c-0ac0-453e-9a95-f55320054ad9-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.681017 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.699525 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p6jmv"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.751849 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerStarted","Data":"7a2426e6cfc347cc1a717c9a966306104dd25d30c2b178309ab403c7304eba33"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.755920 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" event={"ID":"66bcf78c-0ac0-453e-9a95-f55320054ad9","Type":"ContainerDied","Data":"58faa5fe5df7d51a24637ced099efcf45e62284807890d35bc1f1bb1f77da612"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.756075 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bx4mj" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.769382 4972 generic.go:334] "Generic (PLEG): container finished" podID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerID="36b77e94448ed189e8052bb72711769553ff5b99bc7a03ed09bc0687d90e32db" exitCode=0 Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.769533 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" event={"ID":"a43ffba7-dbee-4012-b991-4c809386fcf6","Type":"ContainerDied","Data":"36b77e94448ed189e8052bb72711769553ff5b99bc7a03ed09bc0687d90e32db"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.777714 4972 generic.go:334] "Generic (PLEG): container finished" podID="08476c57-ac49-4644-a9e6-e81340e19370" containerID="563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508" exitCode=0 Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.777810 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" event={"ID":"08476c57-ac49-4644-a9e6-e81340e19370","Type":"ContainerDied","Data":"563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.822710 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874367 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56e07ece-d896-4107-b9b1-6f24b64f7985","Type":"ContainerStarted","Data":"a13a0b81eb05a9eba0d3e7bff399b3a039caf2fec965a2af1068220b4ac666b0"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874422 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nwt8p" event={"ID":"9a71e95a-5289-47ba-aaf9-5574c65d4c7b","Type":"ContainerDied","Data":"f143360d6c5b9d6595c9daaa37318f374c2cb2f135137ca6fc7033dc2ae19d8f"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874451 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44e7c813-beaa-4015-a581-afcaea85077c","Type":"ContainerStarted","Data":"5ce76cd321380aefaf9d0711951b46a077d6e0d8eabf74a05d495c8ca7ac14ac"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874495 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv" event={"ID":"d90126a7-5923-45bd-9e61-29d1e2e3f469","Type":"ContainerStarted","Data":"9162e846aedae9cab02cac78dd27155155456ffdc1e95a8b549eebf978683915"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874508 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerStarted","Data":"aa7446a5a43998a2fbee8e2ee785ebdb631ce8107b26032d4d3066cccc55bdf0"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.874538 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.876905 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bx4mj"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.877787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"daacfde0-2575-4448-a11f-0eb9a1b2a1cb","Type":"ContainerStarted","Data":"7a609c22878b138327aaf45b4b6c4de8a7002dd66a529eff4831fca5d528f4bd"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.881863 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"57d374b4-97c1-4b80-8f4b-95338fe4f385","Type":"ContainerStarted","Data":"b9ead57636b0b5e1d9c2cc3c78bb1db1fdd716575c2db638176981f316a4c67a"} Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.908155 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 28 10:52:37 crc kubenswrapper[4972]: W0228 10:52:37.912305 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ba1065c_4ef3_473c_9576_d673f19a75a1.slice/crio-73c3431e5ff362a4653374d7542a0dc4d32e8dc3e8f31e1093c542e303af6e4f WatchSource:0}: Error finding container 73c3431e5ff362a4653374d7542a0dc4d32e8dc3e8f31e1093c542e303af6e4f: Status 404 returned error can't find the container with id 73c3431e5ff362a4653374d7542a0dc4d32e8dc3e8f31e1093c542e303af6e4f Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.939643 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:37 crc kubenswrapper[4972]: I0228 10:52:37.969020 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwt8p"] Feb 28 10:52:38 crc kubenswrapper[4972]: W0228 10:52:38.505156 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb839e4_8819_45ba_9afd_312cc02a6964.slice/crio-07f16c6532bcd875040122ff27b6d5fffd3777430140a30ecef4f9de70b999d4 WatchSource:0}: Error finding container 07f16c6532bcd875040122ff27b6d5fffd3777430140a30ecef4f9de70b999d4: Status 404 returned error can't find the container with id 07f16c6532bcd875040122ff27b6d5fffd3777430140a30ecef4f9de70b999d4 Feb 28 10:52:38 crc kubenswrapper[4972]: I0228 10:52:38.508392 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5wgfj"] Feb 28 10:52:38 crc kubenswrapper[4972]: I0228 10:52:38.617335 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 28 10:52:38 crc kubenswrapper[4972]: W0228 10:52:38.626869 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec76b1b5_ea80_49cb_ae7c_0190a52640a1.slice/crio-d98832b6a40d965b742ab4d43d718551d2e6f8920f9539926f427696d5343fc4 WatchSource:0}: Error finding container d98832b6a40d965b742ab4d43d718551d2e6f8920f9539926f427696d5343fc4: Status 404 returned error can't find the container with id d98832b6a40d965b742ab4d43d718551d2e6f8920f9539926f427696d5343fc4 Feb 28 10:52:38 crc kubenswrapper[4972]: I0228 10:52:38.892798 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5wgfj" event={"ID":"7eb839e4-8819-45ba-9afd-312cc02a6964","Type":"ContainerStarted","Data":"07f16c6532bcd875040122ff27b6d5fffd3777430140a30ecef4f9de70b999d4"} Feb 28 10:52:38 crc kubenswrapper[4972]: I0228 10:52:38.894914 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec76b1b5-ea80-49cb-ae7c-0190a52640a1","Type":"ContainerStarted","Data":"d98832b6a40d965b742ab4d43d718551d2e6f8920f9539926f427696d5343fc4"} Feb 28 10:52:38 crc kubenswrapper[4972]: I0228 10:52:38.895983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6ba1065c-4ef3-473c-9576-d673f19a75a1","Type":"ContainerStarted","Data":"73c3431e5ff362a4653374d7542a0dc4d32e8dc3e8f31e1093c542e303af6e4f"} Feb 28 10:52:39 crc kubenswrapper[4972]: I0228 10:52:39.804622 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66bcf78c-0ac0-453e-9a95-f55320054ad9" path="/var/lib/kubelet/pods/66bcf78c-0ac0-453e-9a95-f55320054ad9/volumes" Feb 28 10:52:39 crc kubenswrapper[4972]: I0228 10:52:39.806173 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a71e95a-5289-47ba-aaf9-5574c65d4c7b" path="/var/lib/kubelet/pods/9a71e95a-5289-47ba-aaf9-5574c65d4c7b/volumes" Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.924398 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" event={"ID":"08476c57-ac49-4644-a9e6-e81340e19370","Type":"ContainerStarted","Data":"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca"} Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.925085 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.928034 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" event={"ID":"a43ffba7-dbee-4012-b991-4c809386fcf6","Type":"ContainerStarted","Data":"160773ac42a990d107da4199472109cd4f560951aaf32f2bb28eee7fb7dfad94"} Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.929055 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.952945 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" podStartSLOduration=-9223372017.901855 podStartE2EDuration="18.952920422s" podCreationTimestamp="2026-02-28 10:52:22 +0000 UTC" firstStartedPulling="2026-02-28 10:52:22.996387497 +0000 UTC m=+999.908395245" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:52:40.945544288 +0000 UTC m=+1017.857552026" watchObservedRunningTime="2026-02-28 10:52:40.952920422 +0000 UTC m=+1017.864928160" Feb 28 10:52:40 crc kubenswrapper[4972]: I0228 10:52:40.971075 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" podStartSLOduration=5.461710624 podStartE2EDuration="18.971054738s" podCreationTimestamp="2026-02-28 10:52:22 +0000 UTC" firstStartedPulling="2026-02-28 10:52:23.24766354 +0000 UTC m=+1000.159671278" lastFinishedPulling="2026-02-28 10:52:36.757007654 +0000 UTC m=+1013.669015392" observedRunningTime="2026-02-28 10:52:40.967881979 +0000 UTC m=+1017.879889717" watchObservedRunningTime="2026-02-28 10:52:40.971054738 +0000 UTC m=+1017.883062476" Feb 28 10:52:47 crc kubenswrapper[4972]: I0228 10:52:47.408745 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:47 crc kubenswrapper[4972]: I0228 10:52:47.759128 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:52:47 crc kubenswrapper[4972]: I0228 10:52:47.865790 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:48 crc kubenswrapper[4972]: I0228 10:52:48.002618 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="dnsmasq-dns" containerID="cri-o://52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca" gracePeriod=10 Feb 28 10:52:48 crc kubenswrapper[4972]: I0228 10:52:48.842400 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:48 crc kubenswrapper[4972]: I0228 10:52:48.987214 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc\") pod \"08476c57-ac49-4644-a9e6-e81340e19370\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " Feb 28 10:52:48 crc kubenswrapper[4972]: I0228 10:52:48.987622 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wzks\" (UniqueName: \"kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks\") pod \"08476c57-ac49-4644-a9e6-e81340e19370\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " Feb 28 10:52:48 crc kubenswrapper[4972]: I0228 10:52:48.987702 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config\") pod \"08476c57-ac49-4644-a9e6-e81340e19370\" (UID: \"08476c57-ac49-4644-a9e6-e81340e19370\") " Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021424 4972 generic.go:334] "Generic (PLEG): container finished" podID="08476c57-ac49-4644-a9e6-e81340e19370" containerID="52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca" exitCode=0 Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" event={"ID":"08476c57-ac49-4644-a9e6-e81340e19370","Type":"ContainerDied","Data":"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca"} Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021529 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" event={"ID":"08476c57-ac49-4644-a9e6-e81340e19370","Type":"ContainerDied","Data":"f9672a0a03b139bcf6e6d174e0cfd6a8701e9f0be492356f4717c4b763da7d02"} Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021550 4972 scope.go:117] "RemoveContainer" containerID="52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021566 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-f6vp4" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.021656 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks" (OuterVolumeSpecName: "kube-api-access-7wzks") pod "08476c57-ac49-4644-a9e6-e81340e19370" (UID: "08476c57-ac49-4644-a9e6-e81340e19370"). InnerVolumeSpecName "kube-api-access-7wzks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.053299 4972 scope.go:117] "RemoveContainer" containerID="563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.075666 4972 scope.go:117] "RemoveContainer" containerID="52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca" Feb 28 10:52:49 crc kubenswrapper[4972]: E0228 10:52:49.076365 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca\": container with ID starting with 52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca not found: ID does not exist" containerID="52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.076409 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca"} err="failed to get container status \"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca\": rpc error: code = NotFound desc = could not find container \"52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca\": container with ID starting with 52298dfbc79453ef603372d72755d9f22d0d32aa7acbb10f65cc8132200d0eca not found: ID does not exist" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.076438 4972 scope.go:117] "RemoveContainer" containerID="563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508" Feb 28 10:52:49 crc kubenswrapper[4972]: E0228 10:52:49.076781 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508\": container with ID starting with 563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508 not found: ID does not exist" containerID="563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.076811 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508"} err="failed to get container status \"563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508\": rpc error: code = NotFound desc = could not find container \"563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508\": container with ID starting with 563f7648e404a8f7635e503195368983a696a6997cdd65f5b415ed6bbf508508 not found: ID does not exist" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.090268 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wzks\" (UniqueName: \"kubernetes.io/projected/08476c57-ac49-4644-a9e6-e81340e19370-kube-api-access-7wzks\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.549555 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config" (OuterVolumeSpecName: "config") pod "08476c57-ac49-4644-a9e6-e81340e19370" (UID: "08476c57-ac49-4644-a9e6-e81340e19370"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.599222 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.739695 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08476c57-ac49-4644-a9e6-e81340e19370" (UID: "08476c57-ac49-4644-a9e6-e81340e19370"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.802131 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08476c57-ac49-4644-a9e6-e81340e19370-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.946183 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:49 crc kubenswrapper[4972]: I0228 10:52:49.953244 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-f6vp4"] Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.035970 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"56e07ece-d896-4107-b9b1-6f24b64f7985","Type":"ContainerStarted","Data":"bb13753dd0eed115c9a973126728c79d9e7babacbcb79264c82ea8aadae2066f"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.036530 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.038387 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5wgfj" event={"ID":"7eb839e4-8819-45ba-9afd-312cc02a6964","Type":"ContainerStarted","Data":"e5447bad457a4a0a671679e4750c7000d2420438abf84ef4d24dc6774d017b88"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.048427 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec76b1b5-ea80-49cb-ae7c-0190a52640a1","Type":"ContainerStarted","Data":"d1e22c89196751fe656c6d0728d7a54cddaf2992d3268ccad2ea3648a172d8f5"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.053084 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"daacfde0-2575-4448-a11f-0eb9a1b2a1cb","Type":"ContainerStarted","Data":"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.053222 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.060047 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv" event={"ID":"d90126a7-5923-45bd-9e61-29d1e2e3f469","Type":"ContainerStarted","Data":"fd0617b80869d35a84ca60024a713e7d08aba7b182ca53ece6af65c57370e4fe"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.060094 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-p6jmv" Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.069249 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.338407361 podStartE2EDuration="24.069220222s" podCreationTimestamp="2026-02-28 10:52:26 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.465181361 +0000 UTC m=+1014.377189119" lastFinishedPulling="2026-02-28 10:52:47.195994182 +0000 UTC m=+1024.108001980" observedRunningTime="2026-02-28 10:52:50.06054384 +0000 UTC m=+1026.972551578" watchObservedRunningTime="2026-02-28 10:52:50.069220222 +0000 UTC m=+1026.981227970" Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.071277 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6ba1065c-4ef3-473c-9576-d673f19a75a1","Type":"ContainerStarted","Data":"e43f07bd3f32ac5df7a5a46c6befefa89def2343b3cadb723e81f47f8ce723bd"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.086072 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-p6jmv" podStartSLOduration=7.373449527 podStartE2EDuration="17.08604634s" podCreationTimestamp="2026-02-28 10:52:33 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.738636722 +0000 UTC m=+1014.650644460" lastFinishedPulling="2026-02-28 10:52:47.451233535 +0000 UTC m=+1024.363241273" observedRunningTime="2026-02-28 10:52:50.085563336 +0000 UTC m=+1026.997571074" watchObservedRunningTime="2026-02-28 10:52:50.08604634 +0000 UTC m=+1026.998054078" Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.086300 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerStarted","Data":"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.093196 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerStarted","Data":"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.096841 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"57d374b4-97c1-4b80-8f4b-95338fe4f385","Type":"ContainerStarted","Data":"20842070ec96b3ef6317525309e26fd7bfae1229d4f6ac7779d07a0a0bb43d7e"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.100126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44e7c813-beaa-4015-a581-afcaea85077c","Type":"ContainerStarted","Data":"366bbb94d4f4cfe543d825b1d439482fa2cec170aa74d8cca770d4e49ba9656a"} Feb 28 10:52:50 crc kubenswrapper[4972]: I0228 10:52:50.140057 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.438984861 podStartE2EDuration="21.140030402s" podCreationTimestamp="2026-02-28 10:52:29 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.716083684 +0000 UTC m=+1014.628091412" lastFinishedPulling="2026-02-28 10:52:48.417129215 +0000 UTC m=+1025.329136953" observedRunningTime="2026-02-28 10:52:50.129298323 +0000 UTC m=+1027.041306081" watchObservedRunningTime="2026-02-28 10:52:50.140030402 +0000 UTC m=+1027.052038130" Feb 28 10:52:51 crc kubenswrapper[4972]: I0228 10:52:51.114148 4972 generic.go:334] "Generic (PLEG): container finished" podID="7eb839e4-8819-45ba-9afd-312cc02a6964" containerID="e5447bad457a4a0a671679e4750c7000d2420438abf84ef4d24dc6774d017b88" exitCode=0 Feb 28 10:52:51 crc kubenswrapper[4972]: I0228 10:52:51.114387 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5wgfj" event={"ID":"7eb839e4-8819-45ba-9afd-312cc02a6964","Type":"ContainerDied","Data":"e5447bad457a4a0a671679e4750c7000d2420438abf84ef4d24dc6774d017b88"} Feb 28 10:52:51 crc kubenswrapper[4972]: I0228 10:52:51.812206 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08476c57-ac49-4644-a9e6-e81340e19370" path="/var/lib/kubelet/pods/08476c57-ac49-4644-a9e6-e81340e19370/volumes" Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.138777 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ec76b1b5-ea80-49cb-ae7c-0190a52640a1","Type":"ContainerStarted","Data":"a672b8cd565ae893b00b67784adf182a1525beb0b963db21cc22445238402630"} Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.141860 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6ba1065c-4ef3-473c-9576-d673f19a75a1","Type":"ContainerStarted","Data":"edb6d5e7abdc52827cffe9450ed3d2495487cc68d10362aff0c08072bbcbd9d7"} Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.173106 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.950855477 podStartE2EDuration="19.173081739s" podCreationTimestamp="2026-02-28 10:52:34 +0000 UTC" firstStartedPulling="2026-02-28 10:52:38.630842811 +0000 UTC m=+1015.542850569" lastFinishedPulling="2026-02-28 10:52:52.853069093 +0000 UTC m=+1029.765076831" observedRunningTime="2026-02-28 10:52:53.168597175 +0000 UTC m=+1030.080604923" watchObservedRunningTime="2026-02-28 10:52:53.173081739 +0000 UTC m=+1030.085089477" Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.196855 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.306598927 podStartE2EDuration="22.19682165s" podCreationTimestamp="2026-02-28 10:52:31 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.941746884 +0000 UTC m=+1014.853754622" lastFinishedPulling="2026-02-28 10:52:52.831969607 +0000 UTC m=+1029.743977345" observedRunningTime="2026-02-28 10:52:53.191945414 +0000 UTC m=+1030.103953162" watchObservedRunningTime="2026-02-28 10:52:53.19682165 +0000 UTC m=+1030.108829408" Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.666615 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:53 crc kubenswrapper[4972]: I0228 10:52:53.721825 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.159033 4972 generic.go:334] "Generic (PLEG): container finished" podID="44e7c813-beaa-4015-a581-afcaea85077c" containerID="366bbb94d4f4cfe543d825b1d439482fa2cec170aa74d8cca770d4e49ba9656a" exitCode=0 Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.159114 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44e7c813-beaa-4015-a581-afcaea85077c","Type":"ContainerDied","Data":"366bbb94d4f4cfe543d825b1d439482fa2cec170aa74d8cca770d4e49ba9656a"} Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.165191 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5wgfj" event={"ID":"7eb839e4-8819-45ba-9afd-312cc02a6964","Type":"ContainerStarted","Data":"8d60a50dd2f12db23d3b6ae28a72b122e0ea7e6663fd8fbb68fec30180234a70"} Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.165269 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5wgfj" event={"ID":"7eb839e4-8819-45ba-9afd-312cc02a6964","Type":"ContainerStarted","Data":"58f94d0bc5ec40257f8bc67ac44ff53a9e5cb872887e0a036973f93509507f7e"} Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.165485 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.168900 4972 generic.go:334] "Generic (PLEG): container finished" podID="57d374b4-97c1-4b80-8f4b-95338fe4f385" containerID="20842070ec96b3ef6317525309e26fd7bfae1229d4f6ac7779d07a0a0bb43d7e" exitCode=0 Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.168994 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"57d374b4-97c1-4b80-8f4b-95338fe4f385","Type":"ContainerDied","Data":"20842070ec96b3ef6317525309e26fd7bfae1229d4f6ac7779d07a0a0bb43d7e"} Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.170425 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.227199 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5wgfj" podStartSLOduration=12.36331117 podStartE2EDuration="21.227178524s" podCreationTimestamp="2026-02-28 10:52:33 +0000 UTC" firstStartedPulling="2026-02-28 10:52:38.509320369 +0000 UTC m=+1015.421328117" lastFinishedPulling="2026-02-28 10:52:47.373187733 +0000 UTC m=+1024.285195471" observedRunningTime="2026-02-28 10:52:54.222642988 +0000 UTC m=+1031.134650766" watchObservedRunningTime="2026-02-28 10:52:54.227178524 +0000 UTC m=+1031.139186272" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.275482 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.278252 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.329308 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.466409 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:54 crc kubenswrapper[4972]: E0228 10:52:54.471290 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="dnsmasq-dns" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.471316 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="dnsmasq-dns" Feb 28 10:52:54 crc kubenswrapper[4972]: E0228 10:52:54.471341 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="init" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.471349 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="init" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.471524 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="08476c57-ac49-4644-a9e6-e81340e19370" containerName="dnsmasq-dns" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.472511 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.479150 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.484858 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.524174 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-jl8t9"] Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.525270 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.528126 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.554659 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jl8t9"] Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.646958 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647019 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq7jh\" (UniqueName: \"kubernetes.io/projected/0be6aa95-e5d9-4e85-9991-d58945420d69-kube-api-access-vq7jh\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647090 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovn-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647172 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647197 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647217 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-combined-ca-bundle\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647265 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95kn9\" (UniqueName: \"kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647292 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6aa95-e5d9-4e85-9991-d58945420d69-config\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647315 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.647333 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovs-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.749442 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.749661 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq7jh\" (UniqueName: \"kubernetes.io/projected/0be6aa95-e5d9-4e85-9991-d58945420d69-kube-api-access-vq7jh\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.750303 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovn-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.750714 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovn-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.750747 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.750755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751020 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751058 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-combined-ca-bundle\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751133 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95kn9\" (UniqueName: \"kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751168 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6aa95-e5d9-4e85-9991-d58945420d69-config\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751208 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovs-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.751361 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0be6aa95-e5d9-4e85-9991-d58945420d69-ovs-rundir\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.752299 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.752539 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6aa95-e5d9-4e85-9991-d58945420d69-config\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.753057 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.759218 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-combined-ca-bundle\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.767441 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0be6aa95-e5d9-4e85-9991-d58945420d69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.767941 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq7jh\" (UniqueName: \"kubernetes.io/projected/0be6aa95-e5d9-4e85-9991-d58945420d69-kube-api-access-vq7jh\") pod \"ovn-controller-metrics-jl8t9\" (UID: \"0be6aa95-e5d9-4e85-9991-d58945420d69\") " pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.771797 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95kn9\" (UniqueName: \"kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9\") pod \"dnsmasq-dns-7fd796d7df-fzlnq\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.817179 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.858123 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jl8t9" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.878698 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.907868 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.909124 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.912079 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 28 10:52:54 crc kubenswrapper[4972]: I0228 10:52:54.928832 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.063516 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.064185 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.064255 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.064790 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.064954 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qckkb\" (UniqueName: \"kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.169292 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.169385 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qckkb\" (UniqueName: \"kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.169506 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.169637 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.169683 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.170332 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.172199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.173295 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.173258 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.190876 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44e7c813-beaa-4015-a581-afcaea85077c","Type":"ContainerStarted","Data":"160efcc1b8c384a8748a2e70ee2c1ab6ab357b26ebb05d83047471c0c676cfdc"} Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.197257 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qckkb\" (UniqueName: \"kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb\") pod \"dnsmasq-dns-86db49b7ff-pnkbg\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.199142 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"57d374b4-97c1-4b80-8f4b-95338fe4f385","Type":"ContainerStarted","Data":"32a4be68b330f371e8223be9e4efb1d6406f05cfd36805a0a02a87bbc4e187cb"} Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.200004 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.200021 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.213882 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.214322 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.226140 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.269439346 podStartE2EDuration="30.226107603s" podCreationTimestamp="2026-02-28 10:52:25 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.498887639 +0000 UTC m=+1014.410895377" lastFinishedPulling="2026-02-28 10:52:47.455555876 +0000 UTC m=+1024.367563634" observedRunningTime="2026-02-28 10:52:55.221924127 +0000 UTC m=+1032.133931885" watchObservedRunningTime="2026-02-28 10:52:55.226107603 +0000 UTC m=+1032.138115341" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.231162 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.248540 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.255552 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.347855681 podStartE2EDuration="32.255522012s" podCreationTimestamp="2026-02-28 10:52:23 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.46583033 +0000 UTC m=+1014.377838078" lastFinishedPulling="2026-02-28 10:52:47.373496671 +0000 UTC m=+1024.285504409" observedRunningTime="2026-02-28 10:52:55.250011079 +0000 UTC m=+1032.162018817" watchObservedRunningTime="2026-02-28 10:52:55.255522012 +0000 UTC m=+1032.167529750" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.381221 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:55 crc kubenswrapper[4972]: W0228 10:52:55.386655 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda45c8612_507d_4faa_b753_42448dd191e4.slice/crio-b9ff03bedc33e1e9326323edeb7b6dd83a2bddfab8e126ed422fbefde6d413de WatchSource:0}: Error finding container b9ff03bedc33e1e9326323edeb7b6dd83a2bddfab8e126ed422fbefde6d413de: Status 404 returned error can't find the container with id b9ff03bedc33e1e9326323edeb7b6dd83a2bddfab8e126ed422fbefde6d413de Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.440900 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jl8t9"] Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.539609 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.541535 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.545864 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-x6tm4" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.546150 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.546327 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.550837 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.557566 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.696937 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697343 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697391 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9z7p\" (UniqueName: \"kubernetes.io/projected/8ccd0a65-b9b9-4b45-ba11-97903d81f938-kube-api-access-p9z7p\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697429 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-scripts\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697487 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-config\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697506 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.697543 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.798881 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-config\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.798933 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.798982 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.799019 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.799061 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.799097 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9z7p\" (UniqueName: \"kubernetes.io/projected/8ccd0a65-b9b9-4b45-ba11-97903d81f938-kube-api-access-p9z7p\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.799136 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-scripts\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.800220 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-scripts\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.800248 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.800486 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd0a65-b9b9-4b45-ba11-97903d81f938-config\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.805424 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.807066 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.810558 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0a65-b9b9-4b45-ba11-97903d81f938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.831089 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.832614 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9z7p\" (UniqueName: \"kubernetes.io/projected/8ccd0a65-b9b9-4b45-ba11-97903d81f938-kube-api-access-p9z7p\") pod \"ovn-northd-0\" (UID: \"8ccd0a65-b9b9-4b45-ba11-97903d81f938\") " pod="openstack/ovn-northd-0" Feb 28 10:52:55 crc kubenswrapper[4972]: I0228 10:52:55.876397 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.209275 4972 generic.go:334] "Generic (PLEG): container finished" podID="a45c8612-507d-4faa-b753-42448dd191e4" containerID="3977dee12883d73ffcd11415ccf8f1c89d20f137a3fc05d159979989615d0772" exitCode=0 Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.209513 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" event={"ID":"a45c8612-507d-4faa-b753-42448dd191e4","Type":"ContainerDied","Data":"3977dee12883d73ffcd11415ccf8f1c89d20f137a3fc05d159979989615d0772"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.209719 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" event={"ID":"a45c8612-507d-4faa-b753-42448dd191e4","Type":"ContainerStarted","Data":"b9ff03bedc33e1e9326323edeb7b6dd83a2bddfab8e126ed422fbefde6d413de"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.215518 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jl8t9" event={"ID":"0be6aa95-e5d9-4e85-9991-d58945420d69","Type":"ContainerStarted","Data":"fda12f0583eacbfcaf60969d9af74e01f8a0870dd6e3a764c48f01de84a5d973"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.215593 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jl8t9" event={"ID":"0be6aa95-e5d9-4e85-9991-d58945420d69","Type":"ContainerStarted","Data":"ab8108d46b960abeb3b239131e9041fe58fc9f2ddd562417480d08dc2551dc94"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.221975 4972 generic.go:334] "Generic (PLEG): container finished" podID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerID="6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089" exitCode=0 Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.223351 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" event={"ID":"a244d660-2c2b-4270-a2f8-f2a2a665b4ce","Type":"ContainerDied","Data":"6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.223402 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" event={"ID":"a244d660-2c2b-4270-a2f8-f2a2a665b4ce","Type":"ContainerStarted","Data":"1e52315a0252deeea0efbcb9294d7754ecb4561ae323e2be048f975bfbd0c883"} Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.290194 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-jl8t9" podStartSLOduration=2.290162725 podStartE2EDuration="2.290162725s" podCreationTimestamp="2026-02-28 10:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:52:56.281029941 +0000 UTC m=+1033.193037679" watchObservedRunningTime="2026-02-28 10:52:56.290162725 +0000 UTC m=+1033.202170463" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.398778 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.600964 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.652444 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.652552 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.715640 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config\") pod \"a45c8612-507d-4faa-b753-42448dd191e4\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.715719 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95kn9\" (UniqueName: \"kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9\") pod \"a45c8612-507d-4faa-b753-42448dd191e4\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.715752 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc\") pod \"a45c8612-507d-4faa-b753-42448dd191e4\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.715805 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb\") pod \"a45c8612-507d-4faa-b753-42448dd191e4\" (UID: \"a45c8612-507d-4faa-b753-42448dd191e4\") " Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.722301 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9" (OuterVolumeSpecName: "kube-api-access-95kn9") pod "a45c8612-507d-4faa-b753-42448dd191e4" (UID: "a45c8612-507d-4faa-b753-42448dd191e4"). InnerVolumeSpecName "kube-api-access-95kn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.738577 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config" (OuterVolumeSpecName: "config") pod "a45c8612-507d-4faa-b753-42448dd191e4" (UID: "a45c8612-507d-4faa-b753-42448dd191e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.740975 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a45c8612-507d-4faa-b753-42448dd191e4" (UID: "a45c8612-507d-4faa-b753-42448dd191e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.744750 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a45c8612-507d-4faa-b753-42448dd191e4" (UID: "a45c8612-507d-4faa-b753-42448dd191e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.821057 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.821112 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.821133 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95kn9\" (UniqueName: \"kubernetes.io/projected/a45c8612-507d-4faa-b753-42448dd191e4-kube-api-access-95kn9\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.821154 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a45c8612-507d-4faa-b753-42448dd191e4-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:52:56 crc kubenswrapper[4972]: I0228 10:52:56.876559 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.234126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" event={"ID":"a45c8612-507d-4faa-b753-42448dd191e4","Type":"ContainerDied","Data":"b9ff03bedc33e1e9326323edeb7b6dd83a2bddfab8e126ed422fbefde6d413de"} Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.234250 4972 scope.go:117] "RemoveContainer" containerID="3977dee12883d73ffcd11415ccf8f1c89d20f137a3fc05d159979989615d0772" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.234345 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fzlnq" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.237293 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8ccd0a65-b9b9-4b45-ba11-97903d81f938","Type":"ContainerStarted","Data":"d32933c0eeffd03d2da6dd58f917bbc03686335f0a9f476a98e7519e2fa29784"} Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.243943 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" event={"ID":"a244d660-2c2b-4270-a2f8-f2a2a665b4ce","Type":"ContainerStarted","Data":"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced"} Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.244663 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.269046 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" podStartSLOduration=3.269026176 podStartE2EDuration="3.269026176s" podCreationTimestamp="2026-02-28 10:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:52:57.261669151 +0000 UTC m=+1034.173676889" watchObservedRunningTime="2026-02-28 10:52:57.269026176 +0000 UTC m=+1034.181033904" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.325806 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.334104 4972 scope.go:117] "RemoveContainer" containerID="f71e6577840dbe4ba1155be8097374581a4a92b4f8ad1f8585f5b5befce71400" Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.348197 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fzlnq"] Feb 28 10:52:57 crc kubenswrapper[4972]: E0228 10:52:57.438678 4972 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:41926->38.102.83.12:44997: write tcp 38.102.83.12:41926->38.102.83.12:44997: write: broken pipe Feb 28 10:52:57 crc kubenswrapper[4972]: I0228 10:52:57.800866 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a45c8612-507d-4faa-b753-42448dd191e4" path="/var/lib/kubelet/pods/a45c8612-507d-4faa-b753-42448dd191e4/volumes" Feb 28 10:52:58 crc kubenswrapper[4972]: I0228 10:52:58.255689 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8ccd0a65-b9b9-4b45-ba11-97903d81f938","Type":"ContainerStarted","Data":"6c42f27665fae3e542a3a81612cbee328f5790b4e13a0815f4d103d22abe8120"} Feb 28 10:52:58 crc kubenswrapper[4972]: I0228 10:52:58.256186 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8ccd0a65-b9b9-4b45-ba11-97903d81f938","Type":"ContainerStarted","Data":"7874092b171a5274287e78ce771d3aebba29703610fb5f49f7effaba937c088f"} Feb 28 10:52:58 crc kubenswrapper[4972]: I0228 10:52:58.282167 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.065130142 podStartE2EDuration="3.28213857s" podCreationTimestamp="2026-02-28 10:52:55 +0000 UTC" firstStartedPulling="2026-02-28 10:52:56.425535153 +0000 UTC m=+1033.337542891" lastFinishedPulling="2026-02-28 10:52:57.642543581 +0000 UTC m=+1034.554551319" observedRunningTime="2026-02-28 10:52:58.280247898 +0000 UTC m=+1035.192255636" watchObservedRunningTime="2026-02-28 10:52:58.28213857 +0000 UTC m=+1035.194146298" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.263545 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.447050 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.463105 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.511810 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.512171 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="dnsmasq-dns" containerID="cri-o://805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced" gracePeriod=10 Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.563634 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:52:59 crc kubenswrapper[4972]: E0228 10:52:59.564183 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45c8612-507d-4faa-b753-42448dd191e4" containerName="init" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.564205 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45c8612-507d-4faa-b753-42448dd191e4" containerName="init" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.564508 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a45c8612-507d-4faa-b753-42448dd191e4" containerName="init" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.565679 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.642068 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.679593 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.679664 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.679738 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.679762 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj7mn\" (UniqueName: \"kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.679873 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.702083 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.791116 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.791610 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj7mn\" (UniqueName: \"kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.791696 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.791736 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.791760 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.792334 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.792471 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.792983 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.793502 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.822576 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj7mn\" (UniqueName: \"kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn\") pod \"dnsmasq-dns-698758b865-z5slv\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:52:59 crc kubenswrapper[4972]: I0228 10:52:59.986450 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.134593 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.281724 4972 generic.go:334] "Generic (PLEG): container finished" podID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerID="805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced" exitCode=0 Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.283085 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.283180 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" event={"ID":"a244d660-2c2b-4270-a2f8-f2a2a665b4ce","Type":"ContainerDied","Data":"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced"} Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.283231 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-pnkbg" event={"ID":"a244d660-2c2b-4270-a2f8-f2a2a665b4ce","Type":"ContainerDied","Data":"1e52315a0252deeea0efbcb9294d7754ecb4561ae323e2be048f975bfbd0c883"} Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.283257 4972 scope.go:117] "RemoveContainer" containerID="805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.302553 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc\") pod \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.302600 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb\") pod \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.302630 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qckkb\" (UniqueName: \"kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb\") pod \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.302702 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb\") pod \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.303623 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config\") pod \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\" (UID: \"a244d660-2c2b-4270-a2f8-f2a2a665b4ce\") " Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.307389 4972 scope.go:117] "RemoveContainer" containerID="6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.309570 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb" (OuterVolumeSpecName: "kube-api-access-qckkb") pod "a244d660-2c2b-4270-a2f8-f2a2a665b4ce" (UID: "a244d660-2c2b-4270-a2f8-f2a2a665b4ce"). InnerVolumeSpecName "kube-api-access-qckkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.334894 4972 scope.go:117] "RemoveContainer" containerID="805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced" Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.335321 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced\": container with ID starting with 805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced not found: ID does not exist" containerID="805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.335365 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced"} err="failed to get container status \"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced\": rpc error: code = NotFound desc = could not find container \"805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced\": container with ID starting with 805eaf27413a44bfb32d4ccf30f8e0464bdcb49b16b6a10212797ca912bfaced not found: ID does not exist" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.335393 4972 scope.go:117] "RemoveContainer" containerID="6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089" Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.335971 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089\": container with ID starting with 6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089 not found: ID does not exist" containerID="6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.336003 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089"} err="failed to get container status \"6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089\": rpc error: code = NotFound desc = could not find container \"6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089\": container with ID starting with 6c17b11f575c2c7f078cdd892425e14ec2f93faa020fb78107164ab2fb108089 not found: ID does not exist" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.344449 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a244d660-2c2b-4270-a2f8-f2a2a665b4ce" (UID: "a244d660-2c2b-4270-a2f8-f2a2a665b4ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.346108 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a244d660-2c2b-4270-a2f8-f2a2a665b4ce" (UID: "a244d660-2c2b-4270-a2f8-f2a2a665b4ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.349342 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a244d660-2c2b-4270-a2f8-f2a2a665b4ce" (UID: "a244d660-2c2b-4270-a2f8-f2a2a665b4ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.353330 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config" (OuterVolumeSpecName: "config") pod "a244d660-2c2b-4270-a2f8-f2a2a665b4ce" (UID: "a244d660-2c2b-4270-a2f8-f2a2a665b4ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.406901 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.406954 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.406969 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qckkb\" (UniqueName: \"kubernetes.io/projected/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-kube-api-access-qckkb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.406982 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.406992 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a244d660-2c2b-4270-a2f8-f2a2a665b4ce-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.521182 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.635360 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.648380 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-pnkbg"] Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.665667 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.666100 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="init" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.666129 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="init" Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.666159 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="dnsmasq-dns" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.666166 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="dnsmasq-dns" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.666332 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" containerName="dnsmasq-dns" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.671800 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.674254 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-4hmjd" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.674690 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.675111 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.678301 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.690026 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817720 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817778 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-cache\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817836 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-lock\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817865 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhlpw\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-kube-api-access-fhlpw\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817917 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4376e6f8-3008-4aeb-b049-54ee633833f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.817953 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.919977 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.920070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-cache\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.920578 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.920830 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-cache\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.921119 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-lock\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.921484 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4376e6f8-3008-4aeb-b049-54ee633833f2-lock\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.921575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhlpw\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-kube-api-access-fhlpw\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.922002 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4376e6f8-3008-4aeb-b049-54ee633833f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.922594 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.923075 4972 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.923098 4972 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 28 10:53:00 crc kubenswrapper[4972]: E0228 10:53:00.923149 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift podName:4376e6f8-3008-4aeb-b049-54ee633833f2 nodeName:}" failed. No retries permitted until 2026-02-28 10:53:01.423130846 +0000 UTC m=+1038.335138584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift") pod "swift-storage-0" (UID: "4376e6f8-3008-4aeb-b049-54ee633833f2") : configmap "swift-ring-files" not found Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.927741 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4376e6f8-3008-4aeb-b049-54ee633833f2-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.939575 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhlpw\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-kube-api-access-fhlpw\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:00 crc kubenswrapper[4972]: I0228 10:53:00.952015 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.092897 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2f84x"] Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.094272 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.097009 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.103725 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.112111 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.114424 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2f84x"] Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238236 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238312 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238379 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238517 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238599 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm6jv\" (UniqueName: \"kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238665 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.238707 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.301240 4972 generic.go:334] "Generic (PLEG): container finished" podID="130947a9-cb3b-444b-a979-a399900cdcfd" containerID="732a1c0fd9ea3e1fc034064c2eda9289238fc2fc708590418706268bf72d9f06" exitCode=0 Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.301329 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5slv" event={"ID":"130947a9-cb3b-444b-a979-a399900cdcfd","Type":"ContainerDied","Data":"732a1c0fd9ea3e1fc034064c2eda9289238fc2fc708590418706268bf72d9f06"} Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.301363 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5slv" event={"ID":"130947a9-cb3b-444b-a979-a399900cdcfd","Type":"ContainerStarted","Data":"85acdfe8398ee166c684d0022b0cde1838ae04fde2313238df7cc1cb404e518a"} Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.340813 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.340904 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm6jv\" (UniqueName: \"kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.340979 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341007 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341045 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341097 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341900 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.341964 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.342081 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.345452 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.345740 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.345945 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.362300 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm6jv\" (UniqueName: \"kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv\") pod \"swift-ring-rebalance-2f84x\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.377195 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.417083 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.443253 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:01 crc kubenswrapper[4972]: E0228 10:53:01.443573 4972 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 28 10:53:01 crc kubenswrapper[4972]: E0228 10:53:01.443619 4972 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 28 10:53:01 crc kubenswrapper[4972]: E0228 10:53:01.443700 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift podName:4376e6f8-3008-4aeb-b049-54ee633833f2 nodeName:}" failed. No retries permitted until 2026-02-28 10:53:02.443677597 +0000 UTC m=+1039.355685345 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift") pod "swift-storage-0" (UID: "4376e6f8-3008-4aeb-b049-54ee633833f2") : configmap "swift-ring-files" not found Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.485777 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 28 10:53:01 crc kubenswrapper[4972]: I0228 10:53:01.804048 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a244d660-2c2b-4270-a2f8-f2a2a665b4ce" path="/var/lib/kubelet/pods/a244d660-2c2b-4270-a2f8-f2a2a665b4ce/volumes" Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.025182 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2f84x"] Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.312562 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5slv" event={"ID":"130947a9-cb3b-444b-a979-a399900cdcfd","Type":"ContainerStarted","Data":"4e1c7b374e98639acf87411556ad8969f1edb6af427bc3b8d0aa2ff5e108227a"} Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.313028 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.317755 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2f84x" event={"ID":"28171537-6c72-4fe2-9e23-709a2f0824ae","Type":"ContainerStarted","Data":"a1ae2310cd4e547f258d12535883cb91fcb3f4a6ad3dc2acf1b451edd74c26ff"} Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.341512 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-z5slv" podStartSLOduration=3.341484657 podStartE2EDuration="3.341484657s" podCreationTimestamp="2026-02-28 10:52:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:02.334902634 +0000 UTC m=+1039.246910402" watchObservedRunningTime="2026-02-28 10:53:02.341484657 +0000 UTC m=+1039.253492395" Feb 28 10:53:02 crc kubenswrapper[4972]: I0228 10:53:02.496638 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:02 crc kubenswrapper[4972]: E0228 10:53:02.497621 4972 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 28 10:53:02 crc kubenswrapper[4972]: E0228 10:53:02.497662 4972 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 28 10:53:02 crc kubenswrapper[4972]: E0228 10:53:02.497740 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift podName:4376e6f8-3008-4aeb-b049-54ee633833f2 nodeName:}" failed. No retries permitted until 2026-02-28 10:53:04.497710149 +0000 UTC m=+1041.409717887 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift") pod "swift-storage-0" (UID: "4376e6f8-3008-4aeb-b049-54ee633833f2") : configmap "swift-ring-files" not found Feb 28 10:53:03 crc kubenswrapper[4972]: I0228 10:53:03.962842 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-c5qlb"] Feb 28 10:53:03 crc kubenswrapper[4972]: I0228 10:53:03.964238 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:03 crc kubenswrapper[4972]: I0228 10:53:03.967150 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 28 10:53:03 crc kubenswrapper[4972]: I0228 10:53:03.981606 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c5qlb"] Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.132022 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.132630 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhcz\" (UniqueName: \"kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.235777 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhcz\" (UniqueName: \"kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.237294 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.237630 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.266398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhcz\" (UniqueName: \"kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz\") pod \"root-account-create-update-c5qlb\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.289782 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:04 crc kubenswrapper[4972]: I0228 10:53:04.542165 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:04 crc kubenswrapper[4972]: E0228 10:53:04.542498 4972 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 28 10:53:04 crc kubenswrapper[4972]: E0228 10:53:04.542555 4972 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 28 10:53:04 crc kubenswrapper[4972]: E0228 10:53:04.542666 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift podName:4376e6f8-3008-4aeb-b049-54ee633833f2 nodeName:}" failed. No retries permitted until 2026-02-28 10:53:08.542631576 +0000 UTC m=+1045.454639354 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift") pod "swift-storage-0" (UID: "4376e6f8-3008-4aeb-b049-54ee633833f2") : configmap "swift-ring-files" not found Feb 28 10:53:06 crc kubenswrapper[4972]: I0228 10:53:06.316190 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c5qlb"] Feb 28 10:53:06 crc kubenswrapper[4972]: W0228 10:53:06.322790 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode997d572_6c9e_42e5_b782_9e0ae8d9b31f.slice/crio-77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db WatchSource:0}: Error finding container 77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db: Status 404 returned error can't find the container with id 77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db Feb 28 10:53:06 crc kubenswrapper[4972]: I0228 10:53:06.383057 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c5qlb" event={"ID":"e997d572-6c9e-42e5-b782-9e0ae8d9b31f","Type":"ContainerStarted","Data":"77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db"} Feb 28 10:53:06 crc kubenswrapper[4972]: I0228 10:53:06.384697 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2f84x" event={"ID":"28171537-6c72-4fe2-9e23-709a2f0824ae","Type":"ContainerStarted","Data":"237e9fb56a30306713a35568f51af48037210babfc1e758fdfe8002c394de750"} Feb 28 10:53:06 crc kubenswrapper[4972]: I0228 10:53:06.408644 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-2f84x" podStartSLOduration=1.545030358 podStartE2EDuration="5.408620657s" podCreationTimestamp="2026-02-28 10:53:01 +0000 UTC" firstStartedPulling="2026-02-28 10:53:02.040854503 +0000 UTC m=+1038.952862241" lastFinishedPulling="2026-02-28 10:53:05.904444802 +0000 UTC m=+1042.816452540" observedRunningTime="2026-02-28 10:53:06.405913362 +0000 UTC m=+1043.317921110" watchObservedRunningTime="2026-02-28 10:53:06.408620657 +0000 UTC m=+1043.320628415" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.031546 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hnx6d"] Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.032940 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.040039 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hnx6d"] Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.142421 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-cd49-account-create-update-q6sc6"] Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.143635 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.146352 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.155571 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-cd49-account-create-update-q6sc6"] Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.203817 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.203971 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cvj2\" (UniqueName: \"kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.306666 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.307325 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.307474 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cvj2\" (UniqueName: \"kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.307593 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lchp6\" (UniqueName: \"kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.307692 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.330718 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cvj2\" (UniqueName: \"kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2\") pod \"glance-db-create-hnx6d\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.394310 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.398373 4972 generic.go:334] "Generic (PLEG): container finished" podID="e997d572-6c9e-42e5-b782-9e0ae8d9b31f" containerID="cfa79cc5821a935aab1b950283231368dc1cdc570e9ccc09d78463e7e2f02f96" exitCode=0 Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.398496 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c5qlb" event={"ID":"e997d572-6c9e-42e5-b782-9e0ae8d9b31f","Type":"ContainerDied","Data":"cfa79cc5821a935aab1b950283231368dc1cdc570e9ccc09d78463e7e2f02f96"} Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.410214 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lchp6\" (UniqueName: \"kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.411908 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.416128 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.437891 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lchp6\" (UniqueName: \"kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6\") pod \"glance-cd49-account-create-update-q6sc6\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.463266 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.918850 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hnx6d"] Feb 28 10:53:07 crc kubenswrapper[4972]: W0228 10:53:07.922722 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod600c7709_17f1_46b2_9850_673f80b46460.slice/crio-3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26 WatchSource:0}: Error finding container 3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26: Status 404 returned error can't find the container with id 3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26 Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.977789 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g9g7s"] Feb 28 10:53:07 crc kubenswrapper[4972]: I0228 10:53:07.982549 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.015598 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g9g7s"] Feb 28 10:53:08 crc kubenswrapper[4972]: W0228 10:53:08.018697 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8d52cc1_c832_4ae9_929d_06db453ce86d.slice/crio-d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc WatchSource:0}: Error finding container d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc: Status 404 returned error can't find the container with id d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.035640 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-cd49-account-create-update-q6sc6"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.110882 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-55bc-account-create-update-6c2kx"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.112817 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.115653 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.132343 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55bc-account-create-update-6c2kx"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.133093 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.133233 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2vj\" (UniqueName: \"kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.235268 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b75d7\" (UniqueName: \"kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.235350 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2vj\" (UniqueName: \"kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.235449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.235487 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.236673 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.266176 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2vj\" (UniqueName: \"kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj\") pod \"keystone-db-create-g9g7s\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.297484 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-sf2wr"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.298914 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.316291 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.318099 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sf2wr"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.339677 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.339850 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b75d7\" (UniqueName: \"kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.340868 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.362288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b75d7\" (UniqueName: \"kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7\") pod \"keystone-55bc-account-create-update-6c2kx\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.368940 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-36c3-account-create-update-hfvn9"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.371042 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.376127 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.378423 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-36c3-account-create-update-hfvn9"] Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.411115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cd49-account-create-update-q6sc6" event={"ID":"b8d52cc1-c832-4ae9-929d-06db453ce86d","Type":"ContainerStarted","Data":"5bb2e596a390c700583fdc08ea54a8defd23e9f97018d807159e54539a4a53b5"} Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.411178 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cd49-account-create-update-q6sc6" event={"ID":"b8d52cc1-c832-4ae9-929d-06db453ce86d","Type":"ContainerStarted","Data":"d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc"} Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.415014 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hnx6d" event={"ID":"600c7709-17f1-46b2-9850-673f80b46460","Type":"ContainerStarted","Data":"746c616f6ebd943a942c5c492746e775d718d8235863c502d9874665f4e81f57"} Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.415051 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hnx6d" event={"ID":"600c7709-17f1-46b2-9850-673f80b46460","Type":"ContainerStarted","Data":"3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26"} Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.438500 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-cd49-account-create-update-q6sc6" podStartSLOduration=1.438437433 podStartE2EDuration="1.438437433s" podCreationTimestamp="2026-02-28 10:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:08.430934254 +0000 UTC m=+1045.342941992" watchObservedRunningTime="2026-02-28 10:53:08.438437433 +0000 UTC m=+1045.350445181" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.443781 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.443866 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxtjl\" (UniqueName: \"kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.461614 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-hnx6d" podStartSLOduration=1.461573688 podStartE2EDuration="1.461573688s" podCreationTimestamp="2026-02-28 10:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:08.442883937 +0000 UTC m=+1045.354891665" watchObservedRunningTime="2026-02-28 10:53:08.461573688 +0000 UTC m=+1045.373581426" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.545860 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6dfj\" (UniqueName: \"kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.546013 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.546068 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxtjl\" (UniqueName: \"kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.546125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.546155 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: E0228 10:53:08.547022 4972 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 28 10:53:08 crc kubenswrapper[4972]: E0228 10:53:08.547068 4972 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.547118 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: E0228 10:53:08.547152 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift podName:4376e6f8-3008-4aeb-b049-54ee633833f2 nodeName:}" failed. No retries permitted until 2026-02-28 10:53:16.547120391 +0000 UTC m=+1053.459128129 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift") pod "swift-storage-0" (UID: "4376e6f8-3008-4aeb-b049-54ee633833f2") : configmap "swift-ring-files" not found Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.565501 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxtjl\" (UniqueName: \"kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl\") pod \"placement-db-create-sf2wr\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.589660 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.620866 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.652693 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.653014 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6dfj\" (UniqueName: \"kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.654847 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.683845 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6dfj\" (UniqueName: \"kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj\") pod \"placement-36c3-account-create-update-hfvn9\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.706124 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.754071 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.856351 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts\") pod \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.856406 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkhcz\" (UniqueName: \"kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz\") pod \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\" (UID: \"e997d572-6c9e-42e5-b782-9e0ae8d9b31f\") " Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.857518 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e997d572-6c9e-42e5-b782-9e0ae8d9b31f" (UID: "e997d572-6c9e-42e5-b782-9e0ae8d9b31f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.865875 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz" (OuterVolumeSpecName: "kube-api-access-gkhcz") pod "e997d572-6c9e-42e5-b782-9e0ae8d9b31f" (UID: "e997d572-6c9e-42e5-b782-9e0ae8d9b31f"). InnerVolumeSpecName "kube-api-access-gkhcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.959044 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.959089 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkhcz\" (UniqueName: \"kubernetes.io/projected/e997d572-6c9e-42e5-b782-9e0ae8d9b31f-kube-api-access-gkhcz\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:08 crc kubenswrapper[4972]: I0228 10:53:08.980734 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g9g7s"] Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.197834 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55bc-account-create-update-6c2kx"] Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.239738 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sf2wr"] Feb 28 10:53:09 crc kubenswrapper[4972]: W0228 10:53:09.262660 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978045d2_8316_4ee1_b850_56ca45506d54.slice/crio-1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c WatchSource:0}: Error finding container 1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c: Status 404 returned error can't find the container with id 1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c Feb 28 10:53:09 crc kubenswrapper[4972]: W0228 10:53:09.277216 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bc31f7d_a528_4e8b_b7b2_015dce384926.slice/crio-df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea WatchSource:0}: Error finding container df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea: Status 404 returned error can't find the container with id df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.372057 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-36c3-account-create-update-hfvn9"] Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.428057 4972 generic.go:334] "Generic (PLEG): container finished" podID="b8d52cc1-c832-4ae9-929d-06db453ce86d" containerID="5bb2e596a390c700583fdc08ea54a8defd23e9f97018d807159e54539a4a53b5" exitCode=0 Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.428211 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cd49-account-create-update-q6sc6" event={"ID":"b8d52cc1-c832-4ae9-929d-06db453ce86d","Type":"ContainerDied","Data":"5bb2e596a390c700583fdc08ea54a8defd23e9f97018d807159e54539a4a53b5"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.432053 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9g7s" event={"ID":"467c3a6a-1854-4faa-89dd-b925b4790534","Type":"ContainerStarted","Data":"97e1821dce3febd8783d2eedd97f63f026995bfcac3d848f044fc794422f71f7"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.433867 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55bc-account-create-update-6c2kx" event={"ID":"978045d2-8316-4ee1-b850-56ca45506d54","Type":"ContainerStarted","Data":"1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.435208 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sf2wr" event={"ID":"3bc31f7d-a528-4e8b-b7b2-015dce384926","Type":"ContainerStarted","Data":"df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.441378 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-36c3-account-create-update-hfvn9" event={"ID":"f36cf3b0-e583-4304-8e26-3142c89b8dd3","Type":"ContainerStarted","Data":"eac1fca37d76d6eeb8cb85e983c18818fa4799cd7e012ac24ca4983fbafbd472"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.449885 4972 generic.go:334] "Generic (PLEG): container finished" podID="600c7709-17f1-46b2-9850-673f80b46460" containerID="746c616f6ebd943a942c5c492746e775d718d8235863c502d9874665f4e81f57" exitCode=0 Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.449999 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hnx6d" event={"ID":"600c7709-17f1-46b2-9850-673f80b46460","Type":"ContainerDied","Data":"746c616f6ebd943a942c5c492746e775d718d8235863c502d9874665f4e81f57"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.459816 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c5qlb" event={"ID":"e997d572-6c9e-42e5-b782-9e0ae8d9b31f","Type":"ContainerDied","Data":"77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db"} Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.459865 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77befd8263ce9ccc6f99da33363796445d365c8e9499dda950fdd94654dbd1db" Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.459902 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c5qlb" Feb 28 10:53:09 crc kubenswrapper[4972]: I0228 10:53:09.991352 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.062560 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.062935 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="dnsmasq-dns" containerID="cri-o://160773ac42a990d107da4199472109cd4f560951aaf32f2bb28eee7fb7dfad94" gracePeriod=10 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.336111 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-c5qlb"] Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.344976 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-c5qlb"] Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.482775 4972 generic.go:334] "Generic (PLEG): container finished" podID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerID="160773ac42a990d107da4199472109cd4f560951aaf32f2bb28eee7fb7dfad94" exitCode=0 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.482843 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" event={"ID":"a43ffba7-dbee-4012-b991-4c809386fcf6","Type":"ContainerDied","Data":"160773ac42a990d107da4199472109cd4f560951aaf32f2bb28eee7fb7dfad94"} Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.484854 4972 generic.go:334] "Generic (PLEG): container finished" podID="467c3a6a-1854-4faa-89dd-b925b4790534" containerID="e5e7631c598d66ce73b6cc535a6865739386b193898a6494adacbe7486fc0499" exitCode=0 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.484880 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9g7s" event={"ID":"467c3a6a-1854-4faa-89dd-b925b4790534","Type":"ContainerDied","Data":"e5e7631c598d66ce73b6cc535a6865739386b193898a6494adacbe7486fc0499"} Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.488030 4972 generic.go:334] "Generic (PLEG): container finished" podID="978045d2-8316-4ee1-b850-56ca45506d54" containerID="54fc94b3bd71a56f3afd5d3450540a99e13691150c7b35200c05d5246e06f92d" exitCode=0 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.488079 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55bc-account-create-update-6c2kx" event={"ID":"978045d2-8316-4ee1-b850-56ca45506d54","Type":"ContainerDied","Data":"54fc94b3bd71a56f3afd5d3450540a99e13691150c7b35200c05d5246e06f92d"} Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.489787 4972 generic.go:334] "Generic (PLEG): container finished" podID="3bc31f7d-a528-4e8b-b7b2-015dce384926" containerID="5edf0ca16a566aece578fb843af03380f03652784ca7d0361164ca9041a26ef6" exitCode=0 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.489884 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sf2wr" event={"ID":"3bc31f7d-a528-4e8b-b7b2-015dce384926","Type":"ContainerDied","Data":"5edf0ca16a566aece578fb843af03380f03652784ca7d0361164ca9041a26ef6"} Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.492600 4972 generic.go:334] "Generic (PLEG): container finished" podID="f36cf3b0-e583-4304-8e26-3142c89b8dd3" containerID="5f4d45661f3ab9791e14fd2412c98b82bf5e8a8504189159deccd97cdd72a34d" exitCode=0 Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.492654 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-36c3-account-create-update-hfvn9" event={"ID":"f36cf3b0-e583-4304-8e26-3142c89b8dd3","Type":"ContainerDied","Data":"5f4d45661f3ab9791e14fd2412c98b82bf5e8a8504189159deccd97cdd72a34d"} Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.648290 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.810650 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc\") pod \"a43ffba7-dbee-4012-b991-4c809386fcf6\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.811275 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config\") pod \"a43ffba7-dbee-4012-b991-4c809386fcf6\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.811719 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wngkw\" (UniqueName: \"kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw\") pod \"a43ffba7-dbee-4012-b991-4c809386fcf6\" (UID: \"a43ffba7-dbee-4012-b991-4c809386fcf6\") " Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.819251 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw" (OuterVolumeSpecName: "kube-api-access-wngkw") pod "a43ffba7-dbee-4012-b991-4c809386fcf6" (UID: "a43ffba7-dbee-4012-b991-4c809386fcf6"). InnerVolumeSpecName "kube-api-access-wngkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.844965 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.886553 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config" (OuterVolumeSpecName: "config") pod "a43ffba7-dbee-4012-b991-4c809386fcf6" (UID: "a43ffba7-dbee-4012-b991-4c809386fcf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.889113 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a43ffba7-dbee-4012-b991-4c809386fcf6" (UID: "a43ffba7-dbee-4012-b991-4c809386fcf6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.914290 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wngkw\" (UniqueName: \"kubernetes.io/projected/a43ffba7-dbee-4012-b991-4c809386fcf6-kube-api-access-wngkw\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.914330 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:10 crc kubenswrapper[4972]: I0228 10:53:10.914339 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a43ffba7-dbee-4012-b991-4c809386fcf6-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.014972 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts\") pod \"600c7709-17f1-46b2-9850-673f80b46460\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.015028 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cvj2\" (UniqueName: \"kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2\") pod \"600c7709-17f1-46b2-9850-673f80b46460\" (UID: \"600c7709-17f1-46b2-9850-673f80b46460\") " Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.015735 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "600c7709-17f1-46b2-9850-673f80b46460" (UID: "600c7709-17f1-46b2-9850-673f80b46460"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.062200 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.118650 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/600c7709-17f1-46b2-9850-673f80b46460-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.230705 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts\") pod \"b8d52cc1-c832-4ae9-929d-06db453ce86d\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.230876 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lchp6\" (UniqueName: \"kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6\") pod \"b8d52cc1-c832-4ae9-929d-06db453ce86d\" (UID: \"b8d52cc1-c832-4ae9-929d-06db453ce86d\") " Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.231797 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8d52cc1-c832-4ae9-929d-06db453ce86d" (UID: "b8d52cc1-c832-4ae9-929d-06db453ce86d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.234017 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2" (OuterVolumeSpecName: "kube-api-access-8cvj2") pod "600c7709-17f1-46b2-9850-673f80b46460" (UID: "600c7709-17f1-46b2-9850-673f80b46460"). InnerVolumeSpecName "kube-api-access-8cvj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.234923 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6" (OuterVolumeSpecName: "kube-api-access-lchp6") pod "b8d52cc1-c832-4ae9-929d-06db453ce86d" (UID: "b8d52cc1-c832-4ae9-929d-06db453ce86d"). InnerVolumeSpecName "kube-api-access-lchp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.333654 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cvj2\" (UniqueName: \"kubernetes.io/projected/600c7709-17f1-46b2-9850-673f80b46460-kube-api-access-8cvj2\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.333731 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8d52cc1-c832-4ae9-929d-06db453ce86d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.333746 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lchp6\" (UniqueName: \"kubernetes.io/projected/b8d52cc1-c832-4ae9-929d-06db453ce86d-kube-api-access-lchp6\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.503685 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" event={"ID":"a43ffba7-dbee-4012-b991-4c809386fcf6","Type":"ContainerDied","Data":"e274afcd0c606a52efcd1ab0331a8c8de1ccae4ecbfef0b8e52e40a8a6c3a9e5"} Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.503738 4972 scope.go:117] "RemoveContainer" containerID="160773ac42a990d107da4199472109cd4f560951aaf32f2bb28eee7fb7dfad94" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.503876 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wv627" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.514999 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-cd49-account-create-update-q6sc6" event={"ID":"b8d52cc1-c832-4ae9-929d-06db453ce86d","Type":"ContainerDied","Data":"d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc"} Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.515024 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d047b11a2504d11386e7c5afe7a5c1e8f35ab391b733bdf5ab535742b99e8dcc" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.515066 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-cd49-account-create-update-q6sc6" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.524862 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hnx6d" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.532937 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hnx6d" event={"ID":"600c7709-17f1-46b2-9850-673f80b46460","Type":"ContainerDied","Data":"3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26"} Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.532952 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ea16ccb760ba1ea03ff479e1fb67a1b817db07f0a635e7ef3501e72f7501f26" Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.658298 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.659165 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wv627"] Feb 28 10:53:11 crc kubenswrapper[4972]: I0228 10:53:11.666367 4972 scope.go:117] "RemoveContainer" containerID="36b77e94448ed189e8052bb72711769553ff5b99bc7a03ed09bc0687d90e32db" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.025615 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" path="/var/lib/kubelet/pods/a43ffba7-dbee-4012-b991-4c809386fcf6/volumes" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.026629 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e997d572-6c9e-42e5-b782-9e0ae8d9b31f" path="/var/lib/kubelet/pods/e997d572-6c9e-42e5-b782-9e0ae8d9b31f/volumes" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.419901 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jj7n2"] Feb 28 10:53:12 crc kubenswrapper[4972]: E0228 10:53:12.420320 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e997d572-6c9e-42e5-b782-9e0ae8d9b31f" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.420335 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e997d572-6c9e-42e5-b782-9e0ae8d9b31f" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: E0228 10:53:12.420360 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600c7709-17f1-46b2-9850-673f80b46460" containerName="mariadb-database-create" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.420366 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="600c7709-17f1-46b2-9850-673f80b46460" containerName="mariadb-database-create" Feb 28 10:53:12 crc kubenswrapper[4972]: E0228 10:53:12.420382 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="dnsmasq-dns" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.420602 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="dnsmasq-dns" Feb 28 10:53:12 crc kubenswrapper[4972]: E0228 10:53:12.420614 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="init" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.420620 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="init" Feb 28 10:53:12 crc kubenswrapper[4972]: E0228 10:53:12.420633 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d52cc1-c832-4ae9-929d-06db453ce86d" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.420638 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d52cc1-c832-4ae9-929d-06db453ce86d" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.421793 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43ffba7-dbee-4012-b991-4c809386fcf6" containerName="dnsmasq-dns" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.421811 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e997d572-6c9e-42e5-b782-9e0ae8d9b31f" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.421818 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="600c7709-17f1-46b2-9850-673f80b46460" containerName="mariadb-database-create" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.421834 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d52cc1-c832-4ae9-929d-06db453ce86d" containerName="mariadb-account-create-update" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.422647 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.428207 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.441406 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gsngd" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.443333 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.469515 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jj7n2"] Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.517533 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.517587 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mdqd\" (UniqueName: \"kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.517682 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.517729 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.608274 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-36c3-account-create-update-hfvn9" event={"ID":"f36cf3b0-e583-4304-8e26-3142c89b8dd3","Type":"ContainerDied","Data":"eac1fca37d76d6eeb8cb85e983c18818fa4799cd7e012ac24ca4983fbafbd472"} Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.608320 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eac1fca37d76d6eeb8cb85e983c18818fa4799cd7e012ac24ca4983fbafbd472" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.608381 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-36c3-account-create-update-hfvn9" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622204 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6dfj\" (UniqueName: \"kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj\") pod \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622378 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts\") pod \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\" (UID: \"f36cf3b0-e583-4304-8e26-3142c89b8dd3\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622629 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622679 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622715 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.622734 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mdqd\" (UniqueName: \"kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.626815 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f36cf3b0-e583-4304-8e26-3142c89b8dd3" (UID: "f36cf3b0-e583-4304-8e26-3142c89b8dd3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.646808 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.646858 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.647389 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.650687 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj" (OuterVolumeSpecName: "kube-api-access-v6dfj") pod "f36cf3b0-e583-4304-8e26-3142c89b8dd3" (UID: "f36cf3b0-e583-4304-8e26-3142c89b8dd3"). InnerVolumeSpecName "kube-api-access-v6dfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.663279 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mdqd\" (UniqueName: \"kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd\") pod \"glance-db-sync-jj7n2\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.695221 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.697069 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.703490 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.724041 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36cf3b0-e583-4304-8e26-3142c89b8dd3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.724345 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6dfj\" (UniqueName: \"kubernetes.io/projected/f36cf3b0-e583-4304-8e26-3142c89b8dd3-kube-api-access-v6dfj\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.768065 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.825677 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff2vj\" (UniqueName: \"kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj\") pod \"467c3a6a-1854-4faa-89dd-b925b4790534\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.825779 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts\") pod \"3bc31f7d-a528-4e8b-b7b2-015dce384926\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.825876 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxtjl\" (UniqueName: \"kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl\") pod \"3bc31f7d-a528-4e8b-b7b2-015dce384926\" (UID: \"3bc31f7d-a528-4e8b-b7b2-015dce384926\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.825971 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b75d7\" (UniqueName: \"kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7\") pod \"978045d2-8316-4ee1-b850-56ca45506d54\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.826046 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts\") pod \"978045d2-8316-4ee1-b850-56ca45506d54\" (UID: \"978045d2-8316-4ee1-b850-56ca45506d54\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.826087 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts\") pod \"467c3a6a-1854-4faa-89dd-b925b4790534\" (UID: \"467c3a6a-1854-4faa-89dd-b925b4790534\") " Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.827060 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "467c3a6a-1854-4faa-89dd-b925b4790534" (UID: "467c3a6a-1854-4faa-89dd-b925b4790534"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.827139 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "978045d2-8316-4ee1-b850-56ca45506d54" (UID: "978045d2-8316-4ee1-b850-56ca45506d54"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.829127 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3bc31f7d-a528-4e8b-b7b2-015dce384926" (UID: "3bc31f7d-a528-4e8b-b7b2-015dce384926"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.832087 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7" (OuterVolumeSpecName: "kube-api-access-b75d7") pod "978045d2-8316-4ee1-b850-56ca45506d54" (UID: "978045d2-8316-4ee1-b850-56ca45506d54"). InnerVolumeSpecName "kube-api-access-b75d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.832873 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj" (OuterVolumeSpecName: "kube-api-access-ff2vj") pod "467c3a6a-1854-4faa-89dd-b925b4790534" (UID: "467c3a6a-1854-4faa-89dd-b925b4790534"). InnerVolumeSpecName "kube-api-access-ff2vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.834952 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl" (OuterVolumeSpecName: "kube-api-access-hxtjl") pod "3bc31f7d-a528-4e8b-b7b2-015dce384926" (UID: "3bc31f7d-a528-4e8b-b7b2-015dce384926"). InnerVolumeSpecName "kube-api-access-hxtjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928412 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b75d7\" (UniqueName: \"kubernetes.io/projected/978045d2-8316-4ee1-b850-56ca45506d54-kube-api-access-b75d7\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928447 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978045d2-8316-4ee1-b850-56ca45506d54-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928483 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/467c3a6a-1854-4faa-89dd-b925b4790534-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928498 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff2vj\" (UniqueName: \"kubernetes.io/projected/467c3a6a-1854-4faa-89dd-b925b4790534-kube-api-access-ff2vj\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928511 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3bc31f7d-a528-4e8b-b7b2-015dce384926-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:12 crc kubenswrapper[4972]: I0228 10:53:12.928524 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxtjl\" (UniqueName: \"kubernetes.io/projected/3bc31f7d-a528-4e8b-b7b2-015dce384926-kube-api-access-hxtjl\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.376388 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jj7n2"] Feb 28 10:53:13 crc kubenswrapper[4972]: W0228 10:53:13.384365 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15c9ff65_59aa_43d9_8898_781c305625ef.slice/crio-99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c WatchSource:0}: Error finding container 99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c: Status 404 returned error can't find the container with id 99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.621012 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55bc-account-create-update-6c2kx" event={"ID":"978045d2-8316-4ee1-b850-56ca45506d54","Type":"ContainerDied","Data":"1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c"} Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.621565 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1285b3ca133ad083cc2ce5cf3ddca8e9301d8e705b08b785358acecb81836c6c" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.621667 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bc-account-create-update-6c2kx" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.628119 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sf2wr" event={"ID":"3bc31f7d-a528-4e8b-b7b2-015dce384926","Type":"ContainerDied","Data":"df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea"} Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.628180 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df71229d23eb1f75556091a6a590e8deaa8c68d12f748a177b2c398dd7b806ea" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.628254 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sf2wr" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.633545 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jj7n2" event={"ID":"15c9ff65-59aa-43d9-8898-781c305625ef","Type":"ContainerStarted","Data":"99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c"} Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.641183 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g9g7s" event={"ID":"467c3a6a-1854-4faa-89dd-b925b4790534","Type":"ContainerDied","Data":"97e1821dce3febd8783d2eedd97f63f026995bfcac3d848f044fc794422f71f7"} Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.641220 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97e1821dce3febd8783d2eedd97f63f026995bfcac3d848f044fc794422f71f7" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.641284 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g9g7s" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.981930 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-wtb85"] Feb 28 10:53:13 crc kubenswrapper[4972]: E0228 10:53:13.982551 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467c3a6a-1854-4faa-89dd-b925b4790534" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982575 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="467c3a6a-1854-4faa-89dd-b925b4790534" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: E0228 10:53:13.982620 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978045d2-8316-4ee1-b850-56ca45506d54" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982630 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="978045d2-8316-4ee1-b850-56ca45506d54" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: E0228 10:53:13.982664 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc31f7d-a528-4e8b-b7b2-015dce384926" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982675 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc31f7d-a528-4e8b-b7b2-015dce384926" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: E0228 10:53:13.982685 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36cf3b0-e583-4304-8e26-3142c89b8dd3" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982694 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36cf3b0-e583-4304-8e26-3142c89b8dd3" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982900 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="467c3a6a-1854-4faa-89dd-b925b4790534" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982922 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc31f7d-a528-4e8b-b7b2-015dce384926" containerName="mariadb-database-create" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982937 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="978045d2-8316-4ee1-b850-56ca45506d54" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.982951 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36cf3b0-e583-4304-8e26-3142c89b8dd3" containerName="mariadb-account-create-update" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.983749 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.986680 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 28 10:53:13 crc kubenswrapper[4972]: I0228 10:53:13.999193 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-wtb85"] Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.161115 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r798\" (UniqueName: \"kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.161287 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.263081 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.263244 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r798\" (UniqueName: \"kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.263999 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.286358 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r798\" (UniqueName: \"kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798\") pod \"root-account-create-update-wtb85\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.308626 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:14 crc kubenswrapper[4972]: I0228 10:53:14.795301 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-wtb85"] Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.690114 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wtb85" event={"ID":"0bf29e88-65d0-4cf5-8939-a376d341d205","Type":"ContainerStarted","Data":"d14170a389bd3189c90e14bac64852e4ece03ac8764a700c6394d2fd33c6569d"} Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.690477 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wtb85" event={"ID":"0bf29e88-65d0-4cf5-8939-a376d341d205","Type":"ContainerStarted","Data":"54439d474cea0b7f8226ca2d63cd3cf05ffdeb8d0450fc90ae93820481949589"} Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.715793 4972 generic.go:334] "Generic (PLEG): container finished" podID="28171537-6c72-4fe2-9e23-709a2f0824ae" containerID="237e9fb56a30306713a35568f51af48037210babfc1e758fdfe8002c394de750" exitCode=0 Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.715862 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2f84x" event={"ID":"28171537-6c72-4fe2-9e23-709a2f0824ae","Type":"ContainerDied","Data":"237e9fb56a30306713a35568f51af48037210babfc1e758fdfe8002c394de750"} Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.734190 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-wtb85" podStartSLOduration=2.734172851 podStartE2EDuration="2.734172851s" podCreationTimestamp="2026-02-28 10:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:15.72299966 +0000 UTC m=+1052.635007408" watchObservedRunningTime="2026-02-28 10:53:15.734172851 +0000 UTC m=+1052.646180589" Feb 28 10:53:15 crc kubenswrapper[4972]: I0228 10:53:15.933863 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 28 10:53:16 crc kubenswrapper[4972]: I0228 10:53:16.640614 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:16 crc kubenswrapper[4972]: I0228 10:53:16.662384 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4376e6f8-3008-4aeb-b049-54ee633833f2-etc-swift\") pod \"swift-storage-0\" (UID: \"4376e6f8-3008-4aeb-b049-54ee633833f2\") " pod="openstack/swift-storage-0" Feb 28 10:53:16 crc kubenswrapper[4972]: I0228 10:53:16.691902 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 28 10:53:16 crc kubenswrapper[4972]: I0228 10:53:16.729384 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bf29e88-65d0-4cf5-8939-a376d341d205" containerID="d14170a389bd3189c90e14bac64852e4ece03ac8764a700c6394d2fd33c6569d" exitCode=0 Feb 28 10:53:16 crc kubenswrapper[4972]: I0228 10:53:16.729698 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wtb85" event={"ID":"0bf29e88-65d0-4cf5-8939-a376d341d205","Type":"ContainerDied","Data":"d14170a389bd3189c90e14bac64852e4ece03ac8764a700c6394d2fd33c6569d"} Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.183833 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.356818 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.356883 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.356940 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.357094 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.357221 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.357258 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.357309 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm6jv\" (UniqueName: \"kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv\") pod \"28171537-6c72-4fe2-9e23-709a2f0824ae\" (UID: \"28171537-6c72-4fe2-9e23-709a2f0824ae\") " Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.358670 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.359256 4972 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.366224 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv" (OuterVolumeSpecName: "kube-api-access-zm6jv") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "kube-api-access-zm6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.368662 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.391566 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.393442 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.397277 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts" (OuterVolumeSpecName: "scripts") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.398383 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: W0228 10:53:17.399371 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4376e6f8_3008_4aeb_b049_54ee633833f2.slice/crio-98f2709cd71ba3f31f2b25476b6a0aecd6902987e57915cc4bb26243a79fdf8d WatchSource:0}: Error finding container 98f2709cd71ba3f31f2b25476b6a0aecd6902987e57915cc4bb26243a79fdf8d: Status 404 returned error can't find the container with id 98f2709cd71ba3f31f2b25476b6a0aecd6902987e57915cc4bb26243a79fdf8d Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.413229 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "28171537-6c72-4fe2-9e23-709a2f0824ae" (UID: "28171537-6c72-4fe2-9e23-709a2f0824ae"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461389 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461450 4972 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/28171537-6c72-4fe2-9e23-709a2f0824ae-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461493 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm6jv\" (UniqueName: \"kubernetes.io/projected/28171537-6c72-4fe2-9e23-709a2f0824ae-kube-api-access-zm6jv\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461507 4972 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461522 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28171537-6c72-4fe2-9e23-709a2f0824ae-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.461534 4972 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/28171537-6c72-4fe2-9e23-709a2f0824ae-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.741629 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2f84x" event={"ID":"28171537-6c72-4fe2-9e23-709a2f0824ae","Type":"ContainerDied","Data":"a1ae2310cd4e547f258d12535883cb91fcb3f4a6ad3dc2acf1b451edd74c26ff"} Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.741714 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1ae2310cd4e547f258d12535883cb91fcb3f4a6ad3dc2acf1b451edd74c26ff" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.741677 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2f84x" Feb 28 10:53:17 crc kubenswrapper[4972]: I0228 10:53:17.749617 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"98f2709cd71ba3f31f2b25476b6a0aecd6902987e57915cc4bb26243a79fdf8d"} Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.131200 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.278846 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r798\" (UniqueName: \"kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798\") pod \"0bf29e88-65d0-4cf5-8939-a376d341d205\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.278928 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts\") pod \"0bf29e88-65d0-4cf5-8939-a376d341d205\" (UID: \"0bf29e88-65d0-4cf5-8939-a376d341d205\") " Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.280213 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0bf29e88-65d0-4cf5-8939-a376d341d205" (UID: "0bf29e88-65d0-4cf5-8939-a376d341d205"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.286589 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798" (OuterVolumeSpecName: "kube-api-access-5r798") pod "0bf29e88-65d0-4cf5-8939-a376d341d205" (UID: "0bf29e88-65d0-4cf5-8939-a376d341d205"). InnerVolumeSpecName "kube-api-access-5r798". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.384616 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r798\" (UniqueName: \"kubernetes.io/projected/0bf29e88-65d0-4cf5-8939-a376d341d205-kube-api-access-5r798\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.384665 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bf29e88-65d0-4cf5-8939-a376d341d205-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.796000 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wtb85" event={"ID":"0bf29e88-65d0-4cf5-8939-a376d341d205","Type":"ContainerDied","Data":"54439d474cea0b7f8226ca2d63cd3cf05ffdeb8d0450fc90ae93820481949589"} Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.796404 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54439d474cea0b7f8226ca2d63cd3cf05ffdeb8d0450fc90ae93820481949589" Feb 28 10:53:18 crc kubenswrapper[4972]: I0228 10:53:18.796447 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wtb85" Feb 28 10:53:19 crc kubenswrapper[4972]: I0228 10:53:19.808367 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"ce88a71e7e2ce2b67252bc1123444fb51ec7ffc26bfcf7ce8f415e3c059a4a17"} Feb 28 10:53:19 crc kubenswrapper[4972]: I0228 10:53:19.808412 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"a396b630302f48c6b0fee411245550723478313eaa98820ecddf823ce2b7246a"} Feb 28 10:53:19 crc kubenswrapper[4972]: I0228 10:53:19.808422 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"b2ea8ec73d9021891b90020845921f90652384ac398ec29c45e467c1e48aa8b9"} Feb 28 10:53:19 crc kubenswrapper[4972]: I0228 10:53:19.808431 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"4c048ab77a29d81d627c299e697d5d74800a22903f1bb18d2ca9caa4be0b2d42"} Feb 28 10:53:20 crc kubenswrapper[4972]: I0228 10:53:20.308979 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-wtb85"] Feb 28 10:53:20 crc kubenswrapper[4972]: I0228 10:53:20.320655 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-wtb85"] Feb 28 10:53:21 crc kubenswrapper[4972]: I0228 10:53:21.802567 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf29e88-65d0-4cf5-8939-a376d341d205" path="/var/lib/kubelet/pods/0bf29e88-65d0-4cf5-8939-a376d341d205/volumes" Feb 28 10:53:22 crc kubenswrapper[4972]: I0228 10:53:22.835568 4972 generic.go:334] "Generic (PLEG): container finished" podID="ffd40994-f4d6-442c-97fa-6ed417500617" containerID="072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a" exitCode=0 Feb 28 10:53:22 crc kubenswrapper[4972]: I0228 10:53:22.835672 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerDied","Data":"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a"} Feb 28 10:53:22 crc kubenswrapper[4972]: I0228 10:53:22.841751 4972 generic.go:334] "Generic (PLEG): container finished" podID="2d65675a-7f0e-44b3-93df-5aad91061871" containerID="fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b" exitCode=0 Feb 28 10:53:22 crc kubenswrapper[4972]: I0228 10:53:22.841803 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerDied","Data":"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b"} Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.649521 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p6jmv" podUID="d90126a7-5923-45bd-9e61-29d1e2e3f469" containerName="ovn-controller" probeResult="failure" output=< Feb 28 10:53:23 crc kubenswrapper[4972]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 28 10:53:23 crc kubenswrapper[4972]: > Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.744642 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.747374 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5wgfj" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.995329 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p6jmv-config-f558f"] Feb 28 10:53:23 crc kubenswrapper[4972]: E0228 10:53:23.996135 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28171537-6c72-4fe2-9e23-709a2f0824ae" containerName="swift-ring-rebalance" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.996156 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="28171537-6c72-4fe2-9e23-709a2f0824ae" containerName="swift-ring-rebalance" Feb 28 10:53:23 crc kubenswrapper[4972]: E0228 10:53:23.996177 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf29e88-65d0-4cf5-8939-a376d341d205" containerName="mariadb-account-create-update" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.996185 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf29e88-65d0-4cf5-8939-a376d341d205" containerName="mariadb-account-create-update" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.996392 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf29e88-65d0-4cf5-8939-a376d341d205" containerName="mariadb-account-create-update" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.996418 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="28171537-6c72-4fe2-9e23-709a2f0824ae" containerName="swift-ring-rebalance" Feb 28 10:53:23 crc kubenswrapper[4972]: I0228 10:53:23.997160 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.002810 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.017511 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p6jmv-config-f558f"] Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102345 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wz5x\" (UniqueName: \"kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102401 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102428 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102457 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102515 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.102760 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.204862 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wz5x\" (UniqueName: \"kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.204919 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.204942 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.204970 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.205009 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.205069 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.205360 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.205764 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.205809 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.228384 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.238361 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.240704 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wz5x\" (UniqueName: \"kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x\") pod \"ovn-controller-p6jmv-config-f558f\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:24 crc kubenswrapper[4972]: I0228 10:53:24.328923 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.373275 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-5kccl"] Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.378266 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.380505 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.398478 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5kccl"] Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.526695 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.526782 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwth\" (UniqueName: \"kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.627866 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.627942 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwth\" (UniqueName: \"kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.629571 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.655588 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwth\" (UniqueName: \"kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth\") pod \"root-account-create-update-5kccl\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:25 crc kubenswrapper[4972]: I0228 10:53:25.701797 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.411278 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5kccl"] Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.425189 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p6jmv-config-f558f"] Feb 28 10:53:27 crc kubenswrapper[4972]: W0228 10:53:27.510035 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1cad4e7_18d2_4fdb_a545_aa2ff4f540c5.slice/crio-bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f WatchSource:0}: Error finding container bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f: Status 404 returned error can't find the container with id bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f Feb 28 10:53:27 crc kubenswrapper[4972]: W0228 10:53:27.512664 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8be112da_979b_4950_abb5_9d56478eefde.slice/crio-0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef WatchSource:0}: Error finding container 0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef: Status 404 returned error can't find the container with id 0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.907251 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jj7n2" event={"ID":"15c9ff65-59aa-43d9-8898-781c305625ef","Type":"ContainerStarted","Data":"8510522ce294b555b4e04accffd1631d6389767666e8ef127833aad14e6458e6"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.910295 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5kccl" event={"ID":"8be112da-979b-4950-abb5-9d56478eefde","Type":"ContainerStarted","Data":"7b4714b3e7338e8b0db3a5504f19e33365667babe0e2ddb246b544e7d3c969fb"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.910344 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5kccl" event={"ID":"8be112da-979b-4950-abb5-9d56478eefde","Type":"ContainerStarted","Data":"0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.916510 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv-config-f558f" event={"ID":"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5","Type":"ContainerStarted","Data":"23d811b848cda59561d05b9cd5ee90bd85b638ded4a8d0b3e3841f44714313da"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.916715 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv-config-f558f" event={"ID":"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5","Type":"ContainerStarted","Data":"bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.919984 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerStarted","Data":"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.920831 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.927730 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"1283c16280a89a2965dd29140e328a3a6dc80faa321c89b128f77378c49ded38"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.930889 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerStarted","Data":"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09"} Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.931152 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.931740 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jj7n2" podStartSLOduration=2.316069556 podStartE2EDuration="15.931681083s" podCreationTimestamp="2026-02-28 10:53:12 +0000 UTC" firstStartedPulling="2026-02-28 10:53:13.387556521 +0000 UTC m=+1050.299564259" lastFinishedPulling="2026-02-28 10:53:27.003168048 +0000 UTC m=+1063.915175786" observedRunningTime="2026-02-28 10:53:27.927624271 +0000 UTC m=+1064.839632029" watchObservedRunningTime="2026-02-28 10:53:27.931681083 +0000 UTC m=+1064.843688821" Feb 28 10:53:27 crc kubenswrapper[4972]: I0228 10:53:27.958439 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-p6jmv-config-f558f" podStartSLOduration=4.958419988 podStartE2EDuration="4.958419988s" podCreationTimestamp="2026-02-28 10:53:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:27.95453384 +0000 UTC m=+1064.866541588" watchObservedRunningTime="2026-02-28 10:53:27.958419988 +0000 UTC m=+1064.870427726" Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.010223 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-5kccl" podStartSLOduration=3.01019938 podStartE2EDuration="3.01019938s" podCreationTimestamp="2026-02-28 10:53:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:27.992794596 +0000 UTC m=+1064.904802334" watchObservedRunningTime="2026-02-28 10:53:28.01019938 +0000 UTC m=+1064.922207118" Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.029847 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.23986558 podStartE2EDuration="1m6.029824457s" podCreationTimestamp="2026-02-28 10:52:22 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.493208201 +0000 UTC m=+1014.405215939" lastFinishedPulling="2026-02-28 10:52:47.283167078 +0000 UTC m=+1024.195174816" observedRunningTime="2026-02-28 10:53:28.015168359 +0000 UTC m=+1064.927176097" watchObservedRunningTime="2026-02-28 10:53:28.029824457 +0000 UTC m=+1064.941832205" Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.047865 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.321920155 podStartE2EDuration="1m6.04784643s" podCreationTimestamp="2026-02-28 10:52:22 +0000 UTC" firstStartedPulling="2026-02-28 10:52:37.469356298 +0000 UTC m=+1014.381364056" lastFinishedPulling="2026-02-28 10:52:47.195282593 +0000 UTC m=+1024.107290331" observedRunningTime="2026-02-28 10:53:28.038961041 +0000 UTC m=+1064.950968789" watchObservedRunningTime="2026-02-28 10:53:28.04784643 +0000 UTC m=+1064.959854168" Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.654437 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-p6jmv" Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.940021 4972 generic.go:334] "Generic (PLEG): container finished" podID="8be112da-979b-4950-abb5-9d56478eefde" containerID="7b4714b3e7338e8b0db3a5504f19e33365667babe0e2ddb246b544e7d3c969fb" exitCode=0 Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.940084 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5kccl" event={"ID":"8be112da-979b-4950-abb5-9d56478eefde","Type":"ContainerDied","Data":"7b4714b3e7338e8b0db3a5504f19e33365667babe0e2ddb246b544e7d3c969fb"} Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.946997 4972 generic.go:334] "Generic (PLEG): container finished" podID="a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" containerID="23d811b848cda59561d05b9cd5ee90bd85b638ded4a8d0b3e3841f44714313da" exitCode=0 Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.947062 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv-config-f558f" event={"ID":"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5","Type":"ContainerDied","Data":"23d811b848cda59561d05b9cd5ee90bd85b638ded4a8d0b3e3841f44714313da"} Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.951625 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"f52cb91ebd4ba342f55c1d1571bd8f965d49e55a6ea4cfd178a2f25197c3d422"} Feb 28 10:53:28 crc kubenswrapper[4972]: I0228 10:53:28.951669 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"baa523148f689b011eecaeec8af92dc71d7e6dec56ba60b8e6a49fa7fb3bbb3a"} Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.777925 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.782569 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855121 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855236 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855318 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwth\" (UniqueName: \"kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth\") pod \"8be112da-979b-4950-abb5-9d56478eefde\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855352 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855423 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855520 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wz5x\" (UniqueName: \"kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855543 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts\") pod \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\" (UID: \"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.855808 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts\") pod \"8be112da-979b-4950-abb5-9d56478eefde\" (UID: \"8be112da-979b-4950-abb5-9d56478eefde\") " Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.856107 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run" (OuterVolumeSpecName: "var-run") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.856121 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.857203 4972 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.857342 4972 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.857278 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.857321 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8be112da-979b-4950-abb5-9d56478eefde" (UID: "8be112da-979b-4950-abb5-9d56478eefde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.858252 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts" (OuterVolumeSpecName: "scripts") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.858441 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.862952 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x" (OuterVolumeSpecName: "kube-api-access-5wz5x") pod "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" (UID: "a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5"). InnerVolumeSpecName "kube-api-access-5wz5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.863133 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth" (OuterVolumeSpecName: "kube-api-access-pcwth") pod "8be112da-979b-4950-abb5-9d56478eefde" (UID: "8be112da-979b-4950-abb5-9d56478eefde"). InnerVolumeSpecName "kube-api-access-pcwth". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.958761 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwth\" (UniqueName: \"kubernetes.io/projected/8be112da-979b-4950-abb5-9d56478eefde-kube-api-access-pcwth\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.959008 4972 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.959083 4972 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.959141 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wz5x\" (UniqueName: \"kubernetes.io/projected/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-kube-api-access-5wz5x\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.959204 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.959261 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8be112da-979b-4950-abb5-9d56478eefde-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.968272 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p6jmv-config-f558f" event={"ID":"a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5","Type":"ContainerDied","Data":"bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f"} Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.968324 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf6cd596385668ad51264101dffd473d8a57d256335db8e7565efc3f15c09a5f" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.968279 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p6jmv-config-f558f" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.969782 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5kccl" event={"ID":"8be112da-979b-4950-abb5-9d56478eefde","Type":"ContainerDied","Data":"0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef"} Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.969808 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0487376c2142d4c5690d752a2098bc4365775a6b84a423ac3f58b289cada69ef" Feb 28 10:53:30 crc kubenswrapper[4972]: I0228 10:53:30.969942 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5kccl" Feb 28 10:53:31 crc kubenswrapper[4972]: I0228 10:53:31.935382 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p6jmv-config-f558f"] Feb 28 10:53:31 crc kubenswrapper[4972]: I0228 10:53:31.946767 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p6jmv-config-f558f"] Feb 28 10:53:31 crc kubenswrapper[4972]: I0228 10:53:31.985397 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"294312f6507c4878598d7dd1f7951d0b92ff83c039e617e7239cae396b464641"} Feb 28 10:53:33 crc kubenswrapper[4972]: I0228 10:53:33.809638 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" path="/var/lib/kubelet/pods/a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5/volumes" Feb 28 10:53:34 crc kubenswrapper[4972]: I0228 10:53:34.011142 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"5b772ebb72cf6319698d499b99c9210a4c71256f72b939c3bf2bb06c456b4204"} Feb 28 10:53:34 crc kubenswrapper[4972]: I0228 10:53:34.011686 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"caf1e4b30ecca8dede6cab4f8ebaa8423e46e2288964a21fad16f3bda47e44cf"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.027315 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"36148794dd3a0253fe82559f260d7945ef43c7e740fba0ee6300ce9762cf6a76"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.027367 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"a2abaf21b23e5460913df9202e4c11625bd258a65c231c7abc3d619e6a03cb0e"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.027380 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"cb883d99f642e4593e14a2d2703be8a958c22deb5972d75bcbaea8fada7f53ef"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.027389 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"8b82efb62c6f463671b0e797d8a536f7d3c775e84075b1ccc8b3cc0e57a13475"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.027408 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4376e6f8-3008-4aeb-b049-54ee633833f2","Type":"ContainerStarted","Data":"eaac77b81a9c462aa93a9a7e12e02d9cf89660962b8afa4ad6cc105b0e9f5459"} Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.081765 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.064117854 podStartE2EDuration="36.081738805s" podCreationTimestamp="2026-02-28 10:52:59 +0000 UTC" firstStartedPulling="2026-02-28 10:53:17.402042044 +0000 UTC m=+1054.314049802" lastFinishedPulling="2026-02-28 10:53:33.419663015 +0000 UTC m=+1070.331670753" observedRunningTime="2026-02-28 10:53:35.070082751 +0000 UTC m=+1071.982090479" watchObservedRunningTime="2026-02-28 10:53:35.081738805 +0000 UTC m=+1071.993746543" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.364707 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:35 crc kubenswrapper[4972]: E0228 10:53:35.365557 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" containerName="ovn-config" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.365591 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" containerName="ovn-config" Feb 28 10:53:35 crc kubenswrapper[4972]: E0228 10:53:35.365636 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be112da-979b-4950-abb5-9d56478eefde" containerName="mariadb-account-create-update" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.365646 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be112da-979b-4950-abb5-9d56478eefde" containerName="mariadb-account-create-update" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.365867 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be112da-979b-4950-abb5-9d56478eefde" containerName="mariadb-account-create-update" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.365908 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1cad4e7-18d2-4fdb-a545-aa2ff4f540c5" containerName="ovn-config" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.366990 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.372539 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.381331 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.419921 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.419990 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.420018 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.420047 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.420117 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q55th\" (UniqueName: \"kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.420168 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522307 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522433 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522478 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522504 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522558 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.522588 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q55th\" (UniqueName: \"kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.523701 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.523710 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.523769 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.523946 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.524561 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.548882 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q55th\" (UniqueName: \"kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th\") pod \"dnsmasq-dns-764c5664d7-2cwg7\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:35 crc kubenswrapper[4972]: I0228 10:53:35.689899 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:36 crc kubenswrapper[4972]: I0228 10:53:36.176498 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:37 crc kubenswrapper[4972]: I0228 10:53:37.044723 4972 generic.go:334] "Generic (PLEG): container finished" podID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerID="dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e" exitCode=0 Feb 28 10:53:37 crc kubenswrapper[4972]: I0228 10:53:37.044793 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" event={"ID":"3fc7f907-a676-4971-ac2e-ebcd40ceaa50","Type":"ContainerDied","Data":"dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e"} Feb 28 10:53:37 crc kubenswrapper[4972]: I0228 10:53:37.045105 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" event={"ID":"3fc7f907-a676-4971-ac2e-ebcd40ceaa50","Type":"ContainerStarted","Data":"0d48918dc91eacd6105b1d6cf05af45c939134460a26ef4aff43ad40830e3c8d"} Feb 28 10:53:38 crc kubenswrapper[4972]: I0228 10:53:38.054257 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" event={"ID":"3fc7f907-a676-4971-ac2e-ebcd40ceaa50","Type":"ContainerStarted","Data":"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615"} Feb 28 10:53:38 crc kubenswrapper[4972]: I0228 10:53:38.055901 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:38 crc kubenswrapper[4972]: I0228 10:53:38.084201 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" podStartSLOduration=3.084184796 podStartE2EDuration="3.084184796s" podCreationTimestamp="2026-02-28 10:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:38.078020305 +0000 UTC m=+1074.990028043" watchObservedRunningTime="2026-02-28 10:53:38.084184796 +0000 UTC m=+1074.996192534" Feb 28 10:53:39 crc kubenswrapper[4972]: I0228 10:53:39.066213 4972 generic.go:334] "Generic (PLEG): container finished" podID="15c9ff65-59aa-43d9-8898-781c305625ef" containerID="8510522ce294b555b4e04accffd1631d6389767666e8ef127833aad14e6458e6" exitCode=0 Feb 28 10:53:39 crc kubenswrapper[4972]: I0228 10:53:39.066499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jj7n2" event={"ID":"15c9ff65-59aa-43d9-8898-781c305625ef","Type":"ContainerDied","Data":"8510522ce294b555b4e04accffd1631d6389767666e8ef127833aad14e6458e6"} Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.462625 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.617637 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle\") pod \"15c9ff65-59aa-43d9-8898-781c305625ef\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.617729 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data\") pod \"15c9ff65-59aa-43d9-8898-781c305625ef\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.617964 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data\") pod \"15c9ff65-59aa-43d9-8898-781c305625ef\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.618716 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mdqd\" (UniqueName: \"kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd\") pod \"15c9ff65-59aa-43d9-8898-781c305625ef\" (UID: \"15c9ff65-59aa-43d9-8898-781c305625ef\") " Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.629791 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "15c9ff65-59aa-43d9-8898-781c305625ef" (UID: "15c9ff65-59aa-43d9-8898-781c305625ef"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.633851 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd" (OuterVolumeSpecName: "kube-api-access-2mdqd") pod "15c9ff65-59aa-43d9-8898-781c305625ef" (UID: "15c9ff65-59aa-43d9-8898-781c305625ef"). InnerVolumeSpecName "kube-api-access-2mdqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.661569 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15c9ff65-59aa-43d9-8898-781c305625ef" (UID: "15c9ff65-59aa-43d9-8898-781c305625ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.669374 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data" (OuterVolumeSpecName: "config-data") pod "15c9ff65-59aa-43d9-8898-781c305625ef" (UID: "15c9ff65-59aa-43d9-8898-781c305625ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.720402 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.720446 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mdqd\" (UniqueName: \"kubernetes.io/projected/15c9ff65-59aa-43d9-8898-781c305625ef-kube-api-access-2mdqd\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.720475 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:40 crc kubenswrapper[4972]: I0228 10:53:40.720489 4972 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15c9ff65-59aa-43d9-8898-781c305625ef-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.105844 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jj7n2" event={"ID":"15c9ff65-59aa-43d9-8898-781c305625ef","Type":"ContainerDied","Data":"99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c"} Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.105894 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99518cc9a5685262c9234d6b1729ed28466b63ad877c6cb53da2db8caa3dbb7c" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.106008 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jj7n2" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.491251 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.491986 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="dnsmasq-dns" containerID="cri-o://bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615" gracePeriod=10 Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.532595 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:53:41 crc kubenswrapper[4972]: E0228 10:53:41.532974 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c9ff65-59aa-43d9-8898-781c305625ef" containerName="glance-db-sync" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.532991 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c9ff65-59aa-43d9-8898-781c305625ef" containerName="glance-db-sync" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.533142 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c9ff65-59aa-43d9-8898-781c305625ef" containerName="glance-db-sync" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.534026 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.538785 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bllmg\" (UniqueName: \"kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.538882 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.538998 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.539119 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.539204 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.539238 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.564423 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.641751 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.641868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.641962 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.641997 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.642067 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bllmg\" (UniqueName: \"kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.642119 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.643530 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.644234 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.644478 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.644712 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.645893 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.663714 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bllmg\" (UniqueName: \"kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg\") pod \"dnsmasq-dns-74f6bcbc87-tgx4k\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.854990 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:41 crc kubenswrapper[4972]: I0228 10:53:41.971260 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053587 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053708 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053789 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q55th\" (UniqueName: \"kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053811 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053862 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.053887 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config\") pod \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\" (UID: \"3fc7f907-a676-4971-ac2e-ebcd40ceaa50\") " Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.059659 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th" (OuterVolumeSpecName: "kube-api-access-q55th") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "kube-api-access-q55th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.111168 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.116999 4972 generic.go:334] "Generic (PLEG): container finished" podID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerID="bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615" exitCode=0 Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.117047 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" event={"ID":"3fc7f907-a676-4971-ac2e-ebcd40ceaa50","Type":"ContainerDied","Data":"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615"} Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.117131 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" event={"ID":"3fc7f907-a676-4971-ac2e-ebcd40ceaa50","Type":"ContainerDied","Data":"0d48918dc91eacd6105b1d6cf05af45c939134460a26ef4aff43ad40830e3c8d"} Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.117153 4972 scope.go:117] "RemoveContainer" containerID="bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.117332 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-2cwg7" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.122021 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.125172 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config" (OuterVolumeSpecName: "config") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.125200 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.130920 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3fc7f907-a676-4971-ac2e-ebcd40ceaa50" (UID: "3fc7f907-a676-4971-ac2e-ebcd40ceaa50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.153567 4972 scope.go:117] "RemoveContainer" containerID="dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.155518 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.155570 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.155586 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q55th\" (UniqueName: \"kubernetes.io/projected/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-kube-api-access-q55th\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.156628 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.156672 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.156689 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fc7f907-a676-4971-ac2e-ebcd40ceaa50-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.176932 4972 scope.go:117] "RemoveContainer" containerID="bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615" Feb 28 10:53:42 crc kubenswrapper[4972]: E0228 10:53:42.177481 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615\": container with ID starting with bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615 not found: ID does not exist" containerID="bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.177614 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615"} err="failed to get container status \"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615\": rpc error: code = NotFound desc = could not find container \"bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615\": container with ID starting with bd054571c2dc1140822ccb9bee60204ad46442dde30e443d767a389b6fdbd615 not found: ID does not exist" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.177698 4972 scope.go:117] "RemoveContainer" containerID="dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e" Feb 28 10:53:42 crc kubenswrapper[4972]: E0228 10:53:42.178029 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e\": container with ID starting with dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e not found: ID does not exist" containerID="dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.178106 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e"} err="failed to get container status \"dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e\": rpc error: code = NotFound desc = could not find container \"dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e\": container with ID starting with dbeb7a97fdc44b61be24f04c135de1637f4a16f48f16bb351ade3eb009f92d4e not found: ID does not exist" Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.319886 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:53:42 crc kubenswrapper[4972]: W0228 10:53:42.321167 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03debb58_d38a_4470_9274_57f7a61f622f.slice/crio-353609d79ce2681e435d7ea592d1f92e0b28f53ea01c1db3082a7a37b5b033c6 WatchSource:0}: Error finding container 353609d79ce2681e435d7ea592d1f92e0b28f53ea01c1db3082a7a37b5b033c6: Status 404 returned error can't find the container with id 353609d79ce2681e435d7ea592d1f92e0b28f53ea01c1db3082a7a37b5b033c6 Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.464912 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:42 crc kubenswrapper[4972]: I0228 10:53:42.473667 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-2cwg7"] Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.127436 4972 generic.go:334] "Generic (PLEG): container finished" podID="03debb58-d38a-4470-9274-57f7a61f622f" containerID="d153b4cc65055ec3c48638b1ee3c7c1cce21541dabe147c68826e29bbde2a016" exitCode=0 Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.127500 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" event={"ID":"03debb58-d38a-4470-9274-57f7a61f622f","Type":"ContainerDied","Data":"d153b4cc65055ec3c48638b1ee3c7c1cce21541dabe147c68826e29bbde2a016"} Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.127551 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" event={"ID":"03debb58-d38a-4470-9274-57f7a61f622f","Type":"ContainerStarted","Data":"353609d79ce2681e435d7ea592d1f92e0b28f53ea01c1db3082a7a37b5b033c6"} Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.593901 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.826375 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" path="/var/lib/kubelet/pods/3fc7f907-a676-4971-ac2e-ebcd40ceaa50/volumes" Feb 28 10:53:43 crc kubenswrapper[4972]: I0228 10:53:43.917666 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 28 10:53:44 crc kubenswrapper[4972]: I0228 10:53:44.140627 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" event={"ID":"03debb58-d38a-4470-9274-57f7a61f622f","Type":"ContainerStarted","Data":"5f399b212cbb7a69f6e81ec179777bdf6f8e0d3496393cb1ba1a8bd8e02b6345"} Feb 28 10:53:44 crc kubenswrapper[4972]: I0228 10:53:44.140919 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:44 crc kubenswrapper[4972]: I0228 10:53:44.168767 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podStartSLOduration=3.168741917 podStartE2EDuration="3.168741917s" podCreationTimestamp="2026-02-28 10:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:44.166994559 +0000 UTC m=+1081.079002297" watchObservedRunningTime="2026-02-28 10:53:44.168741917 +0000 UTC m=+1081.080749665" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.034725 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wbvzd"] Feb 28 10:53:46 crc kubenswrapper[4972]: E0228 10:53:46.035532 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="init" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.035545 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="init" Feb 28 10:53:46 crc kubenswrapper[4972]: E0228 10:53:46.035575 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="dnsmasq-dns" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.035581 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="dnsmasq-dns" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.035793 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc7f907-a676-4971-ac2e-ebcd40ceaa50" containerName="dnsmasq-dns" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.036332 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.053167 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wbvzd"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.225056 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.225140 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flmdc\" (UniqueName: \"kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.327200 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flmdc\" (UniqueName: \"kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.327387 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.327708 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-r4c8b"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.328394 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.328943 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.354065 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-61a1-account-create-update-xzhhm"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.355396 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.358017 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.361605 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-r4c8b"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.383984 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flmdc\" (UniqueName: \"kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc\") pod \"cinder-db-create-wbvzd\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.419280 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-61a1-account-create-update-xzhhm"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.433612 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2qj7\" (UniqueName: \"kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.433724 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.454754 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-w8vmt"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.456288 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.460820 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.463282 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.463527 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.463631 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9794v" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.468587 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-091e-account-create-update-dhvz5"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.470161 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.473095 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-w8vmt"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.474982 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.523002 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-091e-account-create-update-dhvz5"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.536254 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xqcf\" (UniqueName: \"kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.536322 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2qj7\" (UniqueName: \"kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.536406 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.536499 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.537396 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.562384 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2qj7\" (UniqueName: \"kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7\") pod \"neutron-db-create-r4c8b\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.581608 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d4t8j"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.583659 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.611266 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d4t8j"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638269 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638392 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638443 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jm42\" (UniqueName: \"kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638500 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638534 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxxh6\" (UniqueName: \"kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638561 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.638592 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xqcf\" (UniqueName: \"kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.639916 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.649480 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.660257 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xqcf\" (UniqueName: \"kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf\") pod \"cinder-61a1-account-create-update-xzhhm\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.660845 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.677110 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.739976 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxxh6\" (UniqueName: \"kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740030 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740065 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740085 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8p6v\" (UniqueName: \"kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740145 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740194 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.740229 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jm42\" (UniqueName: \"kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.741256 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.744191 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.746110 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.768702 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c659-account-create-update-xk6mx"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.770379 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c659-account-create-update-xk6mx"] Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.770543 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.777922 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxxh6\" (UniqueName: \"kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6\") pod \"neutron-091e-account-create-update-dhvz5\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.800888 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jm42\" (UniqueName: \"kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42\") pod \"keystone-db-sync-w8vmt\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.807405 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.812979 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.841990 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.842062 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8p6v\" (UniqueName: \"kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.842750 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.856296 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.867890 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8p6v\" (UniqueName: \"kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v\") pod \"barbican-db-create-d4t8j\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.915912 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.944347 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:46 crc kubenswrapper[4972]: I0228 10:53:46.944504 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpjgp\" (UniqueName: \"kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.045873 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.046003 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpjgp\" (UniqueName: \"kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.047291 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.080907 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpjgp\" (UniqueName: \"kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp\") pod \"barbican-c659-account-create-update-xk6mx\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.140750 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:47 crc kubenswrapper[4972]: I0228 10:53:47.397680 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-r4c8b"] Feb 28 10:53:47 crc kubenswrapper[4972]: W0228 10:53:47.402030 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2491fea7_baf2_4aec_bbe6_819143abab82.slice/crio-2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5 WatchSource:0}: Error finding container 2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5: Status 404 returned error can't find the container with id 2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5 Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.528064 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-091e-account-create-update-dhvz5"] Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.723007 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-61a1-account-create-update-xzhhm"] Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.733998 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wbvzd"] Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.744703 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-w8vmt"] Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.754663 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d4t8j"] Feb 28 10:53:48 crc kubenswrapper[4972]: W0228 10:53:47.756790 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb0a233f_72b9_4146_b40c_a4a91620d99f.slice/crio-1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23 WatchSource:0}: Error finding container 1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23: Status 404 returned error can't find the container with id 1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23 Feb 28 10:53:48 crc kubenswrapper[4972]: W0228 10:53:47.776015 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod236d1bdf_c20a_48b4_a1a2_16fd2f47910b.slice/crio-7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64 WatchSource:0}: Error finding container 7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64: Status 404 returned error can't find the container with id 7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64 Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:47.829437 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c659-account-create-update-xk6mx"] Feb 28 10:53:48 crc kubenswrapper[4972]: W0228 10:53:47.855127 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e2760f9_f661_4ec8_8c72_f975c3f21390.slice/crio-823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d WatchSource:0}: Error finding container 823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d: Status 404 returned error can't find the container with id 823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.233020 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wbvzd" event={"ID":"ad6d67a2-7a71-4514-9752-1b319503c1a5","Type":"ContainerStarted","Data":"1fe7677bd53e82f2016a463899133dd11466188d9b258c5dddb71db750f6eaa9"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.233620 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wbvzd" event={"ID":"ad6d67a2-7a71-4514-9752-1b319503c1a5","Type":"ContainerStarted","Data":"e08af4220f5602cc9832c557242468aff6257b24d76ea94f1fc5c51c0affe1d9"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.239147 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c659-account-create-update-xk6mx" event={"ID":"2e2760f9-f661-4ec8-8c72-f975c3f21390","Type":"ContainerStarted","Data":"ecfc1dbe52e293e1adcbfe7fd0859431e4a7bde2179f20328be6a02abba47fc7"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.239207 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c659-account-create-update-xk6mx" event={"ID":"2e2760f9-f661-4ec8-8c72-f975c3f21390","Type":"ContainerStarted","Data":"823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.246958 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-w8vmt" event={"ID":"dab7888a-c6a9-42aa-aac0-0c57d67f8613","Type":"ContainerStarted","Data":"1627b9bcedd3b94dea6a448e6fc762e8f649ddc2debc6babe55ad798c797ebcd"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.252377 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61a1-account-create-update-xzhhm" event={"ID":"eb0a233f-72b9-4146-b40c-a4a91620d99f","Type":"ContainerStarted","Data":"1e5e985b8a443f1d6ce01ab7e6442e29cedf58830c690417ebf19a2a45c7a744"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.252480 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61a1-account-create-update-xzhhm" event={"ID":"eb0a233f-72b9-4146-b40c-a4a91620d99f","Type":"ContainerStarted","Data":"1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.262711 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d4t8j" event={"ID":"236d1bdf-c20a-48b4-a1a2-16fd2f47910b","Type":"ContainerStarted","Data":"ec0cd257659fb427d608ecec5320e080a5326eca92a0d9728b42d849a27d20cf"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.262755 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d4t8j" event={"ID":"236d1bdf-c20a-48b4-a1a2-16fd2f47910b","Type":"ContainerStarted","Data":"7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.270842 4972 generic.go:334] "Generic (PLEG): container finished" podID="2491fea7-baf2-4aec-bbe6-819143abab82" containerID="70f9e4a1543805bd5cca39ced121299e36f9dfd77a29a766b3b68e53ed0c9f6d" exitCode=0 Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.270908 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-r4c8b" event={"ID":"2491fea7-baf2-4aec-bbe6-819143abab82","Type":"ContainerDied","Data":"70f9e4a1543805bd5cca39ced121299e36f9dfd77a29a766b3b68e53ed0c9f6d"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.270932 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-r4c8b" event={"ID":"2491fea7-baf2-4aec-bbe6-819143abab82","Type":"ContainerStarted","Data":"2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.272341 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-091e-account-create-update-dhvz5" event={"ID":"0ac57c57-313c-40e9-8118-8b2bd77be36e","Type":"ContainerStarted","Data":"ad5fac50fe00d317f996a5955886ef02ddafa83beeeb16c0e1fbcb55cf2a4cd1"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.272360 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-091e-account-create-update-dhvz5" event={"ID":"0ac57c57-313c-40e9-8118-8b2bd77be36e","Type":"ContainerStarted","Data":"a6c83d6c24a7f0b279f52e5f041917dd652c129d3eb42ff54565927d0df838e3"} Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.272514 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-wbvzd" podStartSLOduration=3.272486947 podStartE2EDuration="3.272486947s" podCreationTimestamp="2026-02-28 10:53:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:48.26575692 +0000 UTC m=+1085.177764658" watchObservedRunningTime="2026-02-28 10:53:48.272486947 +0000 UTC m=+1085.184494685" Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.361528 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-d4t8j" podStartSLOduration=2.361498867 podStartE2EDuration="2.361498867s" podCreationTimestamp="2026-02-28 10:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:48.356164818 +0000 UTC m=+1085.268172556" watchObservedRunningTime="2026-02-28 10:53:48.361498867 +0000 UTC m=+1085.273506605" Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.363752 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-c659-account-create-update-xk6mx" podStartSLOduration=2.36373636 podStartE2EDuration="2.36373636s" podCreationTimestamp="2026-02-28 10:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:48.292158585 +0000 UTC m=+1085.204166323" watchObservedRunningTime="2026-02-28 10:53:48.36373636 +0000 UTC m=+1085.275744128" Feb 28 10:53:48 crc kubenswrapper[4972]: I0228 10:53:48.397021 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-61a1-account-create-update-xzhhm" podStartSLOduration=2.396982136 podStartE2EDuration="2.396982136s" podCreationTimestamp="2026-02-28 10:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:53:48.376652329 +0000 UTC m=+1085.288660067" watchObservedRunningTime="2026-02-28 10:53:48.396982136 +0000 UTC m=+1085.308989874" Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.282404 4972 generic.go:334] "Generic (PLEG): container finished" podID="eb0a233f-72b9-4146-b40c-a4a91620d99f" containerID="1e5e985b8a443f1d6ce01ab7e6442e29cedf58830c690417ebf19a2a45c7a744" exitCode=0 Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.282552 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61a1-account-create-update-xzhhm" event={"ID":"eb0a233f-72b9-4146-b40c-a4a91620d99f","Type":"ContainerDied","Data":"1e5e985b8a443f1d6ce01ab7e6442e29cedf58830c690417ebf19a2a45c7a744"} Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.285500 4972 generic.go:334] "Generic (PLEG): container finished" podID="236d1bdf-c20a-48b4-a1a2-16fd2f47910b" containerID="ec0cd257659fb427d608ecec5320e080a5326eca92a0d9728b42d849a27d20cf" exitCode=0 Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.285616 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d4t8j" event={"ID":"236d1bdf-c20a-48b4-a1a2-16fd2f47910b","Type":"ContainerDied","Data":"ec0cd257659fb427d608ecec5320e080a5326eca92a0d9728b42d849a27d20cf"} Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.287698 4972 generic.go:334] "Generic (PLEG): container finished" podID="0ac57c57-313c-40e9-8118-8b2bd77be36e" containerID="ad5fac50fe00d317f996a5955886ef02ddafa83beeeb16c0e1fbcb55cf2a4cd1" exitCode=0 Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.287799 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-091e-account-create-update-dhvz5" event={"ID":"0ac57c57-313c-40e9-8118-8b2bd77be36e","Type":"ContainerDied","Data":"ad5fac50fe00d317f996a5955886ef02ddafa83beeeb16c0e1fbcb55cf2a4cd1"} Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.290130 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad6d67a2-7a71-4514-9752-1b319503c1a5" containerID="1fe7677bd53e82f2016a463899133dd11466188d9b258c5dddb71db750f6eaa9" exitCode=0 Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.290190 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wbvzd" event={"ID":"ad6d67a2-7a71-4514-9752-1b319503c1a5","Type":"ContainerDied","Data":"1fe7677bd53e82f2016a463899133dd11466188d9b258c5dddb71db750f6eaa9"} Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.294303 4972 generic.go:334] "Generic (PLEG): container finished" podID="2e2760f9-f661-4ec8-8c72-f975c3f21390" containerID="ecfc1dbe52e293e1adcbfe7fd0859431e4a7bde2179f20328be6a02abba47fc7" exitCode=0 Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.294364 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c659-account-create-update-xk6mx" event={"ID":"2e2760f9-f661-4ec8-8c72-f975c3f21390","Type":"ContainerDied","Data":"ecfc1dbe52e293e1adcbfe7fd0859431e4a7bde2179f20328be6a02abba47fc7"} Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.842916 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.850332 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.909178 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2qj7\" (UniqueName: \"kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7\") pod \"2491fea7-baf2-4aec-bbe6-819143abab82\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.909381 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts\") pod \"2491fea7-baf2-4aec-bbe6-819143abab82\" (UID: \"2491fea7-baf2-4aec-bbe6-819143abab82\") " Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.910314 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2491fea7-baf2-4aec-bbe6-819143abab82" (UID: "2491fea7-baf2-4aec-bbe6-819143abab82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:49 crc kubenswrapper[4972]: I0228 10:53:49.920188 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7" (OuterVolumeSpecName: "kube-api-access-w2qj7") pod "2491fea7-baf2-4aec-bbe6-819143abab82" (UID: "2491fea7-baf2-4aec-bbe6-819143abab82"). InnerVolumeSpecName "kube-api-access-w2qj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.011751 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts\") pod \"0ac57c57-313c-40e9-8118-8b2bd77be36e\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.011924 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxxh6\" (UniqueName: \"kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6\") pod \"0ac57c57-313c-40e9-8118-8b2bd77be36e\" (UID: \"0ac57c57-313c-40e9-8118-8b2bd77be36e\") " Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.012513 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2qj7\" (UniqueName: \"kubernetes.io/projected/2491fea7-baf2-4aec-bbe6-819143abab82-kube-api-access-w2qj7\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.012526 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2491fea7-baf2-4aec-bbe6-819143abab82-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.012646 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ac57c57-313c-40e9-8118-8b2bd77be36e" (UID: "0ac57c57-313c-40e9-8118-8b2bd77be36e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.016348 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6" (OuterVolumeSpecName: "kube-api-access-qxxh6") pod "0ac57c57-313c-40e9-8118-8b2bd77be36e" (UID: "0ac57c57-313c-40e9-8118-8b2bd77be36e"). InnerVolumeSpecName "kube-api-access-qxxh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.114302 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ac57c57-313c-40e9-8118-8b2bd77be36e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.114395 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxxh6\" (UniqueName: \"kubernetes.io/projected/0ac57c57-313c-40e9-8118-8b2bd77be36e-kube-api-access-qxxh6\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.307836 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-r4c8b" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.307874 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-r4c8b" event={"ID":"2491fea7-baf2-4aec-bbe6-819143abab82","Type":"ContainerDied","Data":"2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5"} Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.307971 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d401427076e7155c04b0e15f46babeedc9ec224fbc45fdd0d5519c88c748ae5" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.310622 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-091e-account-create-update-dhvz5" event={"ID":"0ac57c57-313c-40e9-8118-8b2bd77be36e","Type":"ContainerDied","Data":"a6c83d6c24a7f0b279f52e5f041917dd652c129d3eb42ff54565927d0df838e3"} Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.310672 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6c83d6c24a7f0b279f52e5f041917dd652c129d3eb42ff54565927d0df838e3" Feb 28 10:53:50 crc kubenswrapper[4972]: I0228 10:53:50.310690 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-091e-account-create-update-dhvz5" Feb 28 10:53:51 crc kubenswrapper[4972]: I0228 10:53:51.857704 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:53:51 crc kubenswrapper[4972]: I0228 10:53:51.923791 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:53:51 crc kubenswrapper[4972]: I0228 10:53:51.924101 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-z5slv" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="dnsmasq-dns" containerID="cri-o://4e1c7b374e98639acf87411556ad8969f1edb6af427bc3b8d0aa2ff5e108227a" gracePeriod=10 Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.331689 4972 generic.go:334] "Generic (PLEG): container finished" podID="130947a9-cb3b-444b-a979-a399900cdcfd" containerID="4e1c7b374e98639acf87411556ad8969f1edb6af427bc3b8d0aa2ff5e108227a" exitCode=0 Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.331926 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5slv" event={"ID":"130947a9-cb3b-444b-a979-a399900cdcfd","Type":"ContainerDied","Data":"4e1c7b374e98639acf87411556ad8969f1edb6af427bc3b8d0aa2ff5e108227a"} Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.790375 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.794437 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.835853 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.841699 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.881961 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpjgp\" (UniqueName: \"kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp\") pod \"2e2760f9-f661-4ec8-8c72-f975c3f21390\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.882020 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts\") pod \"ad6d67a2-7a71-4514-9752-1b319503c1a5\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.882084 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts\") pod \"2e2760f9-f661-4ec8-8c72-f975c3f21390\" (UID: \"2e2760f9-f661-4ec8-8c72-f975c3f21390\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.882106 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flmdc\" (UniqueName: \"kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc\") pod \"ad6d67a2-7a71-4514-9752-1b319503c1a5\" (UID: \"ad6d67a2-7a71-4514-9752-1b319503c1a5\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.886036 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad6d67a2-7a71-4514-9752-1b319503c1a5" (UID: "ad6d67a2-7a71-4514-9752-1b319503c1a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.886841 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e2760f9-f661-4ec8-8c72-f975c3f21390" (UID: "2e2760f9-f661-4ec8-8c72-f975c3f21390"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.891878 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc" (OuterVolumeSpecName: "kube-api-access-flmdc") pod "ad6d67a2-7a71-4514-9752-1b319503c1a5" (UID: "ad6d67a2-7a71-4514-9752-1b319503c1a5"). InnerVolumeSpecName "kube-api-access-flmdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.906197 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp" (OuterVolumeSpecName: "kube-api-access-tpjgp") pod "2e2760f9-f661-4ec8-8c72-f975c3f21390" (UID: "2e2760f9-f661-4ec8-8c72-f975c3f21390"). InnerVolumeSpecName "kube-api-access-tpjgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.973224 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.985868 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts\") pod \"eb0a233f-72b9-4146-b40c-a4a91620d99f\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.986179 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xqcf\" (UniqueName: \"kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf\") pod \"eb0a233f-72b9-4146-b40c-a4a91620d99f\" (UID: \"eb0a233f-72b9-4146-b40c-a4a91620d99f\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.986209 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts\") pod \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.986241 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8p6v\" (UniqueName: \"kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v\") pod \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\" (UID: \"236d1bdf-c20a-48b4-a1a2-16fd2f47910b\") " Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.986786 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb0a233f-72b9-4146-b40c-a4a91620d99f" (UID: "eb0a233f-72b9-4146-b40c-a4a91620d99f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.986980 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpjgp\" (UniqueName: \"kubernetes.io/projected/2e2760f9-f661-4ec8-8c72-f975c3f21390-kube-api-access-tpjgp\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.987003 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad6d67a2-7a71-4514-9752-1b319503c1a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.987013 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e2760f9-f661-4ec8-8c72-f975c3f21390-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.987026 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flmdc\" (UniqueName: \"kubernetes.io/projected/ad6d67a2-7a71-4514-9752-1b319503c1a5-kube-api-access-flmdc\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.987035 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb0a233f-72b9-4146-b40c-a4a91620d99f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.987497 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "236d1bdf-c20a-48b4-a1a2-16fd2f47910b" (UID: "236d1bdf-c20a-48b4-a1a2-16fd2f47910b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.992013 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf" (OuterVolumeSpecName: "kube-api-access-8xqcf") pod "eb0a233f-72b9-4146-b40c-a4a91620d99f" (UID: "eb0a233f-72b9-4146-b40c-a4a91620d99f"). InnerVolumeSpecName "kube-api-access-8xqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:52 crc kubenswrapper[4972]: I0228 10:53:52.992174 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v" (OuterVolumeSpecName: "kube-api-access-x8p6v") pod "236d1bdf-c20a-48b4-a1a2-16fd2f47910b" (UID: "236d1bdf-c20a-48b4-a1a2-16fd2f47910b"). InnerVolumeSpecName "kube-api-access-x8p6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.088063 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc\") pod \"130947a9-cb3b-444b-a979-a399900cdcfd\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.088359 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config\") pod \"130947a9-cb3b-444b-a979-a399900cdcfd\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.088917 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb\") pod \"130947a9-cb3b-444b-a979-a399900cdcfd\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.089026 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb\") pod \"130947a9-cb3b-444b-a979-a399900cdcfd\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.089116 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj7mn\" (UniqueName: \"kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn\") pod \"130947a9-cb3b-444b-a979-a399900cdcfd\" (UID: \"130947a9-cb3b-444b-a979-a399900cdcfd\") " Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.090669 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xqcf\" (UniqueName: \"kubernetes.io/projected/eb0a233f-72b9-4146-b40c-a4a91620d99f-kube-api-access-8xqcf\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.090697 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.090712 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8p6v\" (UniqueName: \"kubernetes.io/projected/236d1bdf-c20a-48b4-a1a2-16fd2f47910b-kube-api-access-x8p6v\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.093187 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn" (OuterVolumeSpecName: "kube-api-access-mj7mn") pod "130947a9-cb3b-444b-a979-a399900cdcfd" (UID: "130947a9-cb3b-444b-a979-a399900cdcfd"). InnerVolumeSpecName "kube-api-access-mj7mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.141103 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config" (OuterVolumeSpecName: "config") pod "130947a9-cb3b-444b-a979-a399900cdcfd" (UID: "130947a9-cb3b-444b-a979-a399900cdcfd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.145523 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "130947a9-cb3b-444b-a979-a399900cdcfd" (UID: "130947a9-cb3b-444b-a979-a399900cdcfd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.148533 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "130947a9-cb3b-444b-a979-a399900cdcfd" (UID: "130947a9-cb3b-444b-a979-a399900cdcfd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.163958 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "130947a9-cb3b-444b-a979-a399900cdcfd" (UID: "130947a9-cb3b-444b-a979-a399900cdcfd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.192674 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.192717 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.192734 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.192744 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj7mn\" (UniqueName: \"kubernetes.io/projected/130947a9-cb3b-444b-a979-a399900cdcfd-kube-api-access-mj7mn\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.192764 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130947a9-cb3b-444b-a979-a399900cdcfd-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.344321 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d4t8j" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.344622 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d4t8j" event={"ID":"236d1bdf-c20a-48b4-a1a2-16fd2f47910b","Type":"ContainerDied","Data":"7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.344717 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fd3a2752db3579eb6def400fa1832513418e95b5075504cdd5d3e5adf74bd64" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.347753 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5slv" event={"ID":"130947a9-cb3b-444b-a979-a399900cdcfd","Type":"ContainerDied","Data":"85acdfe8398ee166c684d0022b0cde1838ae04fde2313238df7cc1cb404e518a"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.347826 4972 scope.go:117] "RemoveContainer" containerID="4e1c7b374e98639acf87411556ad8969f1edb6af427bc3b8d0aa2ff5e108227a" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.347830 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5slv" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.350088 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wbvzd" event={"ID":"ad6d67a2-7a71-4514-9752-1b319503c1a5","Type":"ContainerDied","Data":"e08af4220f5602cc9832c557242468aff6257b24d76ea94f1fc5c51c0affe1d9"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.350146 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e08af4220f5602cc9832c557242468aff6257b24d76ea94f1fc5c51c0affe1d9" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.350113 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wbvzd" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.352108 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c659-account-create-update-xk6mx" event={"ID":"2e2760f9-f661-4ec8-8c72-f975c3f21390","Type":"ContainerDied","Data":"823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.352155 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="823776193d42715067cee7c80495ff9af5107787f69d5a9e259cc149d1ef389d" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.352219 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c659-account-create-update-xk6mx" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.362690 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-w8vmt" event={"ID":"dab7888a-c6a9-42aa-aac0-0c57d67f8613","Type":"ContainerStarted","Data":"ff9df8f9887f1c72ffc34ff7165bbe807ceae4ff7ded130002b308c5548457f6"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.378559 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61a1-account-create-update-xzhhm" event={"ID":"eb0a233f-72b9-4146-b40c-a4a91620d99f","Type":"ContainerDied","Data":"1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23"} Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.378615 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3477f06e3233a66734973e77b00a7f2e92b12c293bbdcebfa25e5b113d6f23" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.378633 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61a1-account-create-update-xzhhm" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.383210 4972 scope.go:117] "RemoveContainer" containerID="732a1c0fd9ea3e1fc034064c2eda9289238fc2fc708590418706268bf72d9f06" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.413399 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-w8vmt" podStartSLOduration=2.533503338 podStartE2EDuration="7.413374809s" podCreationTimestamp="2026-02-28 10:53:46 +0000 UTC" firstStartedPulling="2026-02-28 10:53:47.757028408 +0000 UTC m=+1084.669036146" lastFinishedPulling="2026-02-28 10:53:52.636899879 +0000 UTC m=+1089.548907617" observedRunningTime="2026-02-28 10:53:53.39580657 +0000 UTC m=+1090.307814308" watchObservedRunningTime="2026-02-28 10:53:53.413374809 +0000 UTC m=+1090.325382547" Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.421634 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.431308 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5slv"] Feb 28 10:53:53 crc kubenswrapper[4972]: I0228 10:53:53.837089 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" path="/var/lib/kubelet/pods/130947a9-cb3b-444b-a979-a399900cdcfd/volumes" Feb 28 10:53:56 crc kubenswrapper[4972]: I0228 10:53:56.408837 4972 generic.go:334] "Generic (PLEG): container finished" podID="dab7888a-c6a9-42aa-aac0-0c57d67f8613" containerID="ff9df8f9887f1c72ffc34ff7165bbe807ceae4ff7ded130002b308c5548457f6" exitCode=0 Feb 28 10:53:56 crc kubenswrapper[4972]: I0228 10:53:56.408888 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-w8vmt" event={"ID":"dab7888a-c6a9-42aa-aac0-0c57d67f8613","Type":"ContainerDied","Data":"ff9df8f9887f1c72ffc34ff7165bbe807ceae4ff7ded130002b308c5548457f6"} Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.770792 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.898553 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle\") pod \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.899307 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jm42\" (UniqueName: \"kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42\") pod \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.899420 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data\") pod \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\" (UID: \"dab7888a-c6a9-42aa-aac0-0c57d67f8613\") " Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.906129 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42" (OuterVolumeSpecName: "kube-api-access-8jm42") pod "dab7888a-c6a9-42aa-aac0-0c57d67f8613" (UID: "dab7888a-c6a9-42aa-aac0-0c57d67f8613"). InnerVolumeSpecName "kube-api-access-8jm42". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.927709 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dab7888a-c6a9-42aa-aac0-0c57d67f8613" (UID: "dab7888a-c6a9-42aa-aac0-0c57d67f8613"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:57 crc kubenswrapper[4972]: I0228 10:53:57.955691 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data" (OuterVolumeSpecName: "config-data") pod "dab7888a-c6a9-42aa-aac0-0c57d67f8613" (UID: "dab7888a-c6a9-42aa-aac0-0c57d67f8613"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.002114 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.002171 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jm42\" (UniqueName: \"kubernetes.io/projected/dab7888a-c6a9-42aa-aac0-0c57d67f8613-kube-api-access-8jm42\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.002188 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dab7888a-c6a9-42aa-aac0-0c57d67f8613-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.427139 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-w8vmt" event={"ID":"dab7888a-c6a9-42aa-aac0-0c57d67f8613","Type":"ContainerDied","Data":"1627b9bcedd3b94dea6a448e6fc762e8f649ddc2debc6babe55ad798c797ebcd"} Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.427400 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1627b9bcedd3b94dea6a448e6fc762e8f649ddc2debc6babe55ad798c797ebcd" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.427252 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-w8vmt" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682225 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682682 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6d67a2-7a71-4514-9752-1b319503c1a5" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682699 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6d67a2-7a71-4514-9752-1b319503c1a5" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682720 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2491fea7-baf2-4aec-bbe6-819143abab82" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682728 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2491fea7-baf2-4aec-bbe6-819143abab82" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682739 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="init" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682747 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="init" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682760 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236d1bdf-c20a-48b4-a1a2-16fd2f47910b" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682767 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="236d1bdf-c20a-48b4-a1a2-16fd2f47910b" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682778 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab7888a-c6a9-42aa-aac0-0c57d67f8613" containerName="keystone-db-sync" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682786 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab7888a-c6a9-42aa-aac0-0c57d67f8613" containerName="keystone-db-sync" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682796 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac57c57-313c-40e9-8118-8b2bd77be36e" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682803 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac57c57-313c-40e9-8118-8b2bd77be36e" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682827 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="dnsmasq-dns" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682834 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="dnsmasq-dns" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682863 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb0a233f-72b9-4146-b40c-a4a91620d99f" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682871 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb0a233f-72b9-4146-b40c-a4a91620d99f" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: E0228 10:53:58.682893 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2760f9-f661-4ec8-8c72-f975c3f21390" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.682903 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2760f9-f661-4ec8-8c72-f975c3f21390" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683114 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="236d1bdf-c20a-48b4-a1a2-16fd2f47910b" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683132 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb0a233f-72b9-4146-b40c-a4a91620d99f" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683141 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab7888a-c6a9-42aa-aac0-0c57d67f8613" containerName="keystone-db-sync" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683151 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6d67a2-7a71-4514-9752-1b319503c1a5" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683160 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac57c57-313c-40e9-8118-8b2bd77be36e" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683168 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2760f9-f661-4ec8-8c72-f975c3f21390" containerName="mariadb-account-create-update" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683180 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="130947a9-cb3b-444b-a979-a399900cdcfd" containerName="dnsmasq-dns" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.683198 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2491fea7-baf2-4aec-bbe6-819143abab82" containerName="mariadb-database-create" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.684317 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.717950 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.795835 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vkvhp"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.797593 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.803296 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.803347 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.803621 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.806805 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.809101 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9794v" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.813860 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vkvhp"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828037 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828245 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828383 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828435 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twq48\" (UniqueName: \"kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828547 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.828614 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.892499 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.892585 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.918337 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-llqn9"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.924742 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.928280 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.928580 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.929954 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.929998 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930055 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930076 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930095 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twq48\" (UniqueName: \"kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930125 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930143 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930163 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930195 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930225 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930252 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.930268 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntdxq\" (UniqueName: \"kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.931273 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.931826 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.932763 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.932841 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.932848 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.936630 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rm7g4" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.959235 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-llqn9"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.974192 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.976271 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.985939 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.986268 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.986423 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-b4pp9" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.986876 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twq48\" (UniqueName: \"kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48\") pod \"dnsmasq-dns-847c4cc679-wkrzc\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.989561 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 28 10:53:58 crc kubenswrapper[4972]: I0228 10:53:58.998109 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.014929 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032650 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032720 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032795 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032838 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032889 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032935 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25ds5\" (UniqueName: \"kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.032976 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.033004 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntdxq\" (UniqueName: \"kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.033041 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.033078 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.033109 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.048899 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.058246 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.058766 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.065241 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.070233 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.076038 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntdxq\" (UniqueName: \"kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq\") pod \"keystone-bootstrap-vkvhp\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136685 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136758 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136794 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25ds5\" (UniqueName: \"kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136831 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136860 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136903 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7f7c\" (UniqueName: \"kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136936 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136959 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.136984 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.137006 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.137031 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.137273 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.145886 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.152686 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.166247 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.166715 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.171699 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25ds5\" (UniqueName: \"kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.180301 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data\") pod \"cinder-db-sync-llqn9\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.198534 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vppxs"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.199756 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.213538 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.213808 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.214546 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jdxf6" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.221441 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qtm2c"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.227056 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.229997 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.237755 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244580 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7f7c\" (UniqueName: \"kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244638 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244664 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244706 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244734 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.244844 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k9jst" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.245888 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.247923 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.253648 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-llqn9" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.255203 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.280378 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.283179 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.301404 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.316645 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.319745 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vppxs"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.320509 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.322206 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7f7c\" (UniqueName: \"kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c\") pod \"horizon-54f8c57c9c-kq5wx\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355304 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355602 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355657 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355760 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdsmf\" (UniqueName: \"kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355832 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrtws\" (UniqueName: \"kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.355859 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.359430 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qtm2c"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.389624 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.446206 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.478746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdsmf\" (UniqueName: \"kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479345 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrtws\" (UniqueName: \"kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479395 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479447 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479652 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479740 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479877 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.479930 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.481775 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.481838 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.481936 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.482011 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5tlh\" (UniqueName: \"kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.482071 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.485982 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.490206 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.493600 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.498377 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.503158 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.512205 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.524406 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrtws\" (UniqueName: \"kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws\") pod \"barbican-db-sync-qtm2c\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.532840 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdsmf\" (UniqueName: \"kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf\") pod \"neutron-db-sync-vppxs\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.546011 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.552259 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601271 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601380 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601425 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601468 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601531 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601572 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5tlh\" (UniqueName: \"kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601600 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.601862 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vppxs" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.603261 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.604398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.613130 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4kmth"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.636526 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.637798 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.642260 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.642640 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.647143 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wp8gv" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.647177 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.648905 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.652447 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5tlh\" (UniqueName: \"kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.658363 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.667952 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.680876 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.699807 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.723932 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.724452 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.727075 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.727181 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.727274 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2kkm\" (UniqueName: \"kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.727350 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728178 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728322 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728376 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728401 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728424 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxrzk\" (UniqueName: \"kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.728604 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.768292 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4kmth"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.818610 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.822897 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.823025 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.827666 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.827760 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.827960 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.828060 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gsngd" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.837898 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.837946 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.837971 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxrzk\" (UniqueName: \"kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.838038 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.838078 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.838098 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.840746 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.841284 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.841431 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.841453 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842040 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842142 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842280 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2kkm\" (UniqueName: \"kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842310 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842346 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zzxx\" (UniqueName: \"kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842398 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842437 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842617 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.842771 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.845892 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.847123 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.847222 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.847869 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.850312 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.851168 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.878774 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxrzk\" (UniqueName: \"kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk\") pod \"horizon-5db7884c8f-qvwtf\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.901189 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2kkm\" (UniqueName: \"kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm\") pod \"dnsmasq-dns-785d8bcb8c-clq4v\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.949887 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957036 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957087 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zzxx\" (UniqueName: \"kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957122 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgnf\" (UniqueName: \"kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957148 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957174 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957227 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957252 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957290 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957309 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957336 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957355 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957378 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.957416 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.958239 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.967226 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.970902 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.972160 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.983490 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zzxx\" (UniqueName: \"kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx\") pod \"placement-db-sync-4kmth\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " pod="openstack/placement-db-sync-4kmth" Feb 28 10:53:59 crc kubenswrapper[4972]: I0228 10:53:59.993635 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.041240 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.058574 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kmth" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059135 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgnf\" (UniqueName: \"kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059167 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059192 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059254 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059273 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059312 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059331 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059357 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.059683 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.066596 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.067958 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.071276 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.072239 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.073370 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.087136 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgnf\" (UniqueName: \"kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.090575 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.102944 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.161755 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.164138 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.166020 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.174183 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.175064 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.183071 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537934-2kjkh"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.186093 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.192862 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.193187 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.193419 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.216141 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.255320 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537934-2kjkh"] Feb 28 10:54:00 crc kubenswrapper[4972]: W0228 10:54:00.256971 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53609995_2a99_40b6_8ef7_bbeb7f20eb89.slice/crio-e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1 WatchSource:0}: Error finding container e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1: Status 404 returned error can't find the container with id e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1 Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.262446 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.262646 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.262775 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.262866 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.262938 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcbgb\" (UniqueName: \"kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb\") pod \"auto-csr-approver-29537934-2kjkh\" (UID: \"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3\") " pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.263346 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.263400 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.263477 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.263524 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58d29\" (UniqueName: \"kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: W0228 10:54:00.298065 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedd35ce5_02ce_4909_89c5_6dfce27e50a1.slice/crio-5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09 WatchSource:0}: Error finding container 5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09: Status 404 returned error can't find the container with id 5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09 Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.325960 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vkvhp"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.365755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367488 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367540 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcbgb\" (UniqueName: \"kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb\") pod \"auto-csr-approver-29537934-2kjkh\" (UID: \"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3\") " pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367704 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367730 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367766 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367790 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58d29\" (UniqueName: \"kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367820 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.367867 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.371893 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.374288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.376950 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.379602 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.380742 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.390567 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.392302 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-llqn9"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.402718 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.427409 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.432437 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcbgb\" (UniqueName: \"kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb\") pod \"auto-csr-approver-29537934-2kjkh\" (UID: \"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3\") " pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.438656 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58d29\" (UniqueName: \"kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.443196 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.506855 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vkvhp" event={"ID":"53609995-2a99-40b6-8ef7-bbeb7f20eb89","Type":"ContainerStarted","Data":"e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1"} Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.508018 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-llqn9" event={"ID":"edd35ce5-02ce-4909-89c5-6dfce27e50a1","Type":"ContainerStarted","Data":"5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09"} Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.511441 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" event={"ID":"dc9235e2-1a92-40b5-925a-a33a5ddb9099","Type":"ContainerStarted","Data":"28ef99819f2c85b8e6d9c8ebc1f7bb25c71ac987ba46b23e99fbfc2304b28664"} Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.518583 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54f8c57c9c-kq5wx" event={"ID":"ae5e2238-8b10-439d-a815-610ee23d23fc","Type":"ContainerStarted","Data":"e2c04469f9d5e8b08b68dc56839f5b7cc141867d48a7aa793b91a8e78e0ada3a"} Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.562822 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vppxs"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.563983 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.582434 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.583164 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.605043 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qtm2c"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.624332 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:54:00 crc kubenswrapper[4972]: W0228 10:54:00.625449 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84e043f8_8906_4baf_a610_53446e9c0f85.slice/crio-944e6d267bdb266a5a395de936b90fc9553f9b8d9ec3e1898f8995ae3d9c5abd WatchSource:0}: Error finding container 944e6d267bdb266a5a395de936b90fc9553f9b8d9ec3e1898f8995ae3d9c5abd: Status 404 returned error can't find the container with id 944e6d267bdb266a5a395de936b90fc9553f9b8d9ec3e1898f8995ae3d9c5abd Feb 28 10:54:00 crc kubenswrapper[4972]: W0228 10:54:00.633903 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96039bc0_559a_464e_bc9c_f79fba169a2b.slice/crio-2e0da091dcf8dbbb9d5d1926ea645aa94c5c997d8c6a15e1af45e11667ce6200 WatchSource:0}: Error finding container 2e0da091dcf8dbbb9d5d1926ea645aa94c5c997d8c6a15e1af45e11667ce6200: Status 404 returned error can't find the container with id 2e0da091dcf8dbbb9d5d1926ea645aa94c5c997d8c6a15e1af45e11667ce6200 Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.903446 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:54:00 crc kubenswrapper[4972]: I0228 10:54:00.924006 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4kmth"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.148551 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.271385 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.373929 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.382651 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.403748 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.403808 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcsdq\" (UniqueName: \"kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.403846 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.403948 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.403974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.408018 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.449068 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.503629 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537934-2kjkh"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.505976 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.506021 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcsdq\" (UniqueName: \"kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.506054 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.506148 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.506181 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.507491 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.507738 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.511566 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.512965 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.527916 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.577801 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.607132 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcsdq\" (UniqueName: \"kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq\") pod \"horizon-6d76fdd467-knmck\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.616724 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerStarted","Data":"944e6d267bdb266a5a395de936b90fc9553f9b8d9ec3e1898f8995ae3d9c5abd"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.645223 4972 generic.go:334] "Generic (PLEG): container finished" podID="dc9235e2-1a92-40b5-925a-a33a5ddb9099" containerID="4c812cc3109978cd20a50981cda3c8251a6c2212cddfd666c1e3d87c0562dfaa" exitCode=0 Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.645556 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" event={"ID":"dc9235e2-1a92-40b5-925a-a33a5ddb9099","Type":"ContainerDied","Data":"4c812cc3109978cd20a50981cda3c8251a6c2212cddfd666c1e3d87c0562dfaa"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.665837 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kmth" event={"ID":"7dc233b8-4780-4e84-a0b6-dd5fe2289e16","Type":"ContainerStarted","Data":"c86efac22b4a5c2c031a278c8f9a6ad8190d167c7444dd2a470fbf5f85adfe34"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.679059 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtm2c" event={"ID":"aa1673c6-2e2d-496f-b152-ca9df91df6d7","Type":"ContainerStarted","Data":"22ecb3ffeb62e6b9938c8245ef2c211afd9d9665f95b2a6e070c4db7eaac3b23"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.682784 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:01 crc kubenswrapper[4972]: W0228 10:54:01.715721 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb92a57f4_00aa_404f_8263_b0c8046c6811.slice/crio-1924fba11ac268c737dc798aebe984ec9768bebefbd4d0cde9bb1b18dd350b97 WatchSource:0}: Error finding container 1924fba11ac268c737dc798aebe984ec9768bebefbd4d0cde9bb1b18dd350b97: Status 404 returned error can't find the container with id 1924fba11ac268c737dc798aebe984ec9768bebefbd4d0cde9bb1b18dd350b97 Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.727901 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" event={"ID":"a622dc9d-f76e-4927-8dde-b0d9dd83220e","Type":"ContainerStarted","Data":"83b3a3ab8337334c40a871f3625192868a99b8cb6c4419a8cd595bd705311fd5"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.731355 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.888474 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vppxs" podStartSLOduration=2.888429913 podStartE2EDuration="2.888429913s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:01.842597816 +0000 UTC m=+1098.754605564" watchObservedRunningTime="2026-02-28 10:54:01.888429913 +0000 UTC m=+1098.800437651" Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.897616 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vppxs" event={"ID":"ef9e4339-b034-40f0-a1f0-04229463c846","Type":"ContainerStarted","Data":"39215e36b6acf94dfdcb4712e55fdad5b1116ed4b55c3f2229e2c570a7079184"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.897677 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vppxs" event={"ID":"ef9e4339-b034-40f0-a1f0-04229463c846","Type":"ContainerStarted","Data":"924c6704501cbaf038d4b6be13d2b23aa1d4f5084955fe278590b962e5ecd346"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.897687 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5db7884c8f-qvwtf" event={"ID":"96039bc0-559a-464e-bc9c-f79fba169a2b","Type":"ContainerStarted","Data":"2e0da091dcf8dbbb9d5d1926ea645aa94c5c997d8c6a15e1af45e11667ce6200"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.897699 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vkvhp" event={"ID":"53609995-2a99-40b6-8ef7-bbeb7f20eb89","Type":"ContainerStarted","Data":"8be33d659d45e2374c60706378eb0db0668124a28ee7b3a2e70f6638aa5acd42"} Feb 28 10:54:01 crc kubenswrapper[4972]: I0228 10:54:01.964004 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vkvhp" podStartSLOduration=3.963979288 podStartE2EDuration="3.963979288s" podCreationTimestamp="2026-02-28 10:53:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:01.929995 +0000 UTC m=+1098.842002738" watchObservedRunningTime="2026-02-28 10:54:01.963979288 +0000 UTC m=+1098.875987026" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.336735 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.452059 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.452288 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.453302 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twq48\" (UniqueName: \"kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.453452 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.453560 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.453686 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0\") pod \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\" (UID: \"dc9235e2-1a92-40b5-925a-a33a5ddb9099\") " Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.501513 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.523173 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config" (OuterVolumeSpecName: "config") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.524101 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.534997 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.536830 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.538870 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48" (OuterVolumeSpecName: "kube-api-access-twq48") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "kube-api-access-twq48". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.556747 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.557177 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.557221 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twq48\" (UniqueName: \"kubernetes.io/projected/dc9235e2-1a92-40b5-925a-a33a5ddb9099-kube-api-access-twq48\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.557235 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.557251 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.557264 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.571413 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc9235e2-1a92-40b5-925a-a33a5ddb9099" (UID: "dc9235e2-1a92-40b5-925a-a33a5ddb9099"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.658748 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc9235e2-1a92-40b5-925a-a33a5ddb9099-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.915060 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerStarted","Data":"c26dec33e35e74b5f7b178f48c332501c41be9988288e21f1009590a15e17b68"} Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.917493 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d76fdd467-knmck" event={"ID":"f20522e2-5d4f-45d9-92e0-c51892b78a41","Type":"ContainerStarted","Data":"4450cf1348dfd8006a6f71e720a82c2631f06f410c23627a8ab84a3a07c19c2b"} Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.922631 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" event={"ID":"dc9235e2-1a92-40b5-925a-a33a5ddb9099","Type":"ContainerDied","Data":"28ef99819f2c85b8e6d9c8ebc1f7bb25c71ac987ba46b23e99fbfc2304b28664"} Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.922684 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-wkrzc" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.922714 4972 scope.go:117] "RemoveContainer" containerID="4c812cc3109978cd20a50981cda3c8251a6c2212cddfd666c1e3d87c0562dfaa" Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.931635 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" event={"ID":"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3","Type":"ContainerStarted","Data":"bc167efc253c6fcbe0b41df7f2cb4e387c011b8f111025655a9529e53c9d9059"} Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.937824 4972 generic.go:334] "Generic (PLEG): container finished" podID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerID="afe1436935f6510da9991562f2efe2c1dd01c9f1be8ff0bddf9da2d9fcba1683" exitCode=0 Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.937962 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" event={"ID":"a622dc9d-f76e-4927-8dde-b0d9dd83220e","Type":"ContainerDied","Data":"afe1436935f6510da9991562f2efe2c1dd01c9f1be8ff0bddf9da2d9fcba1683"} Feb 28 10:54:02 crc kubenswrapper[4972]: I0228 10:54:02.960545 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerStarted","Data":"1924fba11ac268c737dc798aebe984ec9768bebefbd4d0cde9bb1b18dd350b97"} Feb 28 10:54:03 crc kubenswrapper[4972]: I0228 10:54:03.040267 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:54:03 crc kubenswrapper[4972]: I0228 10:54:03.051705 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-wkrzc"] Feb 28 10:54:03 crc kubenswrapper[4972]: I0228 10:54:03.834928 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc9235e2-1a92-40b5-925a-a33a5ddb9099" path="/var/lib/kubelet/pods/dc9235e2-1a92-40b5-925a-a33a5ddb9099/volumes" Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.077806 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" event={"ID":"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3","Type":"ContainerStarted","Data":"b1d5898eac1c4fad64bd70d62c1af61c4d98753f97bd6ca2da2729ce3c6b7021"} Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.119825 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" event={"ID":"a622dc9d-f76e-4927-8dde-b0d9dd83220e","Type":"ContainerStarted","Data":"4c1f954e8b237c827ce1f2f63600d7eae87ec64c0c2f97136851e301b8e05e55"} Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.120710 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.131787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerStarted","Data":"e2a3ff4c70fff3c0486ff02e44988c3ef586bd11d5539f999aa6aabf02cc0f3a"} Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.152498 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerStarted","Data":"2bca29557d677209fa721dbc218c9fc183d20bd48f31eddc244b8dcddc8809f2"} Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.188107 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" podStartSLOduration=3.218472264 podStartE2EDuration="4.188087545s" podCreationTimestamp="2026-02-28 10:54:00 +0000 UTC" firstStartedPulling="2026-02-28 10:54:01.615075228 +0000 UTC m=+1098.527082966" lastFinishedPulling="2026-02-28 10:54:02.584690509 +0000 UTC m=+1099.496698247" observedRunningTime="2026-02-28 10:54:04.165592509 +0000 UTC m=+1101.077600247" watchObservedRunningTime="2026-02-28 10:54:04.188087545 +0000 UTC m=+1101.100095283" Feb 28 10:54:04 crc kubenswrapper[4972]: I0228 10:54:04.200235 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" podStartSLOduration=5.200214234 podStartE2EDuration="5.200214234s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:04.192357225 +0000 UTC m=+1101.104364963" watchObservedRunningTime="2026-02-28 10:54:04.200214234 +0000 UTC m=+1101.112221972" Feb 28 10:54:05 crc kubenswrapper[4972]: I0228 10:54:05.175407 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerStarted","Data":"4ba5d0491a86ef873e6bf6b3c66dd4e5e40b87fe90a545af21ee4343e786469c"} Feb 28 10:54:05 crc kubenswrapper[4972]: I0228 10:54:05.177885 4972 generic.go:334] "Generic (PLEG): container finished" podID="625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" containerID="b1d5898eac1c4fad64bd70d62c1af61c4d98753f97bd6ca2da2729ce3c6b7021" exitCode=0 Feb 28 10:54:05 crc kubenswrapper[4972]: I0228 10:54:05.178186 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" event={"ID":"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3","Type":"ContainerDied","Data":"b1d5898eac1c4fad64bd70d62c1af61c4d98753f97bd6ca2da2729ce3c6b7021"} Feb 28 10:54:05 crc kubenswrapper[4972]: I0228 10:54:05.182954 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerStarted","Data":"c6348203524815c0eaf13d9c19ebc28ef73b33576c027466ab98c35bc0fd53c2"} Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.196158 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-log" containerID="cri-o://e2a3ff4c70fff3c0486ff02e44988c3ef586bd11d5539f999aa6aabf02cc0f3a" gracePeriod=30 Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.196209 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-httpd" containerID="cri-o://4ba5d0491a86ef873e6bf6b3c66dd4e5e40b87fe90a545af21ee4343e786469c" gracePeriod=30 Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.196172 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-log" containerID="cri-o://2bca29557d677209fa721dbc218c9fc183d20bd48f31eddc244b8dcddc8809f2" gracePeriod=30 Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.196332 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-httpd" containerID="cri-o://c6348203524815c0eaf13d9c19ebc28ef73b33576c027466ab98c35bc0fd53c2" gracePeriod=30 Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.230718 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.230694078 podStartE2EDuration="7.230694078s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:06.223569329 +0000 UTC m=+1103.135577067" watchObservedRunningTime="2026-02-28 10:54:06.230694078 +0000 UTC m=+1103.142701806" Feb 28 10:54:06 crc kubenswrapper[4972]: I0228 10:54:06.256567 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.256549338 podStartE2EDuration="7.256549338s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:06.247409593 +0000 UTC m=+1103.159417331" watchObservedRunningTime="2026-02-28 10:54:06.256549338 +0000 UTC m=+1103.168557076" Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.213552 4972 generic.go:334] "Generic (PLEG): container finished" podID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerID="c6348203524815c0eaf13d9c19ebc28ef73b33576c027466ab98c35bc0fd53c2" exitCode=0 Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.213895 4972 generic.go:334] "Generic (PLEG): container finished" podID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerID="e2a3ff4c70fff3c0486ff02e44988c3ef586bd11d5539f999aa6aabf02cc0f3a" exitCode=143 Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.213703 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerDied","Data":"c6348203524815c0eaf13d9c19ebc28ef73b33576c027466ab98c35bc0fd53c2"} Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.213969 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerDied","Data":"e2a3ff4c70fff3c0486ff02e44988c3ef586bd11d5539f999aa6aabf02cc0f3a"} Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.216723 4972 generic.go:334] "Generic (PLEG): container finished" podID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerID="4ba5d0491a86ef873e6bf6b3c66dd4e5e40b87fe90a545af21ee4343e786469c" exitCode=0 Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.216769 4972 generic.go:334] "Generic (PLEG): container finished" podID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerID="2bca29557d677209fa721dbc218c9fc183d20bd48f31eddc244b8dcddc8809f2" exitCode=143 Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.216776 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerDied","Data":"4ba5d0491a86ef873e6bf6b3c66dd4e5e40b87fe90a545af21ee4343e786469c"} Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.216806 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerDied","Data":"2bca29557d677209fa721dbc218c9fc183d20bd48f31eddc244b8dcddc8809f2"} Feb 28 10:54:07 crc kubenswrapper[4972]: I0228 10:54:07.971520 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.007914 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:54:08 crc kubenswrapper[4972]: E0228 10:54:08.008628 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9235e2-1a92-40b5-925a-a33a5ddb9099" containerName="init" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.008663 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9235e2-1a92-40b5-925a-a33a5ddb9099" containerName="init" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.008948 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc9235e2-1a92-40b5-925a-a33a5ddb9099" containerName="init" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.010552 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.022973 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027213 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027401 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027589 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027700 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027817 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.027940 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmvvq\" (UniqueName: \"kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.028044 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.037186 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.105342 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129628 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129705 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129730 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129788 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmvvq\" (UniqueName: \"kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129821 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129855 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.129877 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.131368 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.134437 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.134443 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.157199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.157378 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmvvq\" (UniqueName: \"kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.157790 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.165526 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c6cf8f456-722rs"] Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.171671 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.175175 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key\") pod \"horizon-7c898cfcfb-6nt9v\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.190396 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c6cf8f456-722rs"] Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.268764 4972 generic.go:334] "Generic (PLEG): container finished" podID="53609995-2a99-40b6-8ef7-bbeb7f20eb89" containerID="8be33d659d45e2374c60706378eb0db0668124a28ee7b3a2e70f6638aa5acd42" exitCode=0 Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.268818 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vkvhp" event={"ID":"53609995-2a99-40b6-8ef7-bbeb7f20eb89","Type":"ContainerDied","Data":"8be33d659d45e2374c60706378eb0db0668124a28ee7b3a2e70f6638aa5acd42"} Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.351575 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-config-data\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.351750 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-tls-certs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.351792 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-logs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.351987 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-combined-ca-bundle\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.352177 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-scripts\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.352270 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-secret-key\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.352324 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzz5h\" (UniqueName: \"kubernetes.io/projected/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-kube-api-access-gzz5h\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.375500 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.454430 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-logs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.454960 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-combined-ca-bundle\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455037 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-scripts\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455087 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-secret-key\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455099 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-logs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455122 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzz5h\" (UniqueName: \"kubernetes.io/projected/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-kube-api-access-gzz5h\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455680 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-config-data\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.455794 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-tls-certs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.456254 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-scripts\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.456936 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-config-data\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.464152 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-secret-key\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.480004 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-horizon-tls-certs\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.484054 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-combined-ca-bundle\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.487144 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzz5h\" (UniqueName: \"kubernetes.io/projected/a24fd1fd-763d-4106-87b0-805e2f1fc6c5-kube-api-access-gzz5h\") pod \"horizon-7c6cf8f456-722rs\" (UID: \"a24fd1fd-763d-4106-87b0-805e2f1fc6c5\") " pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:08 crc kubenswrapper[4972]: I0228 10:54:08.554673 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:10 crc kubenswrapper[4972]: I0228 10:54:10.044747 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:54:10 crc kubenswrapper[4972]: I0228 10:54:10.117835 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:54:10 crc kubenswrapper[4972]: I0228 10:54:10.118524 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" containerID="cri-o://5f399b212cbb7a69f6e81ec179777bdf6f8e0d3496393cb1ba1a8bd8e02b6345" gracePeriod=10 Feb 28 10:54:10 crc kubenswrapper[4972]: I0228 10:54:10.296047 4972 generic.go:334] "Generic (PLEG): container finished" podID="03debb58-d38a-4470-9274-57f7a61f622f" containerID="5f399b212cbb7a69f6e81ec179777bdf6f8e0d3496393cb1ba1a8bd8e02b6345" exitCode=0 Feb 28 10:54:10 crc kubenswrapper[4972]: I0228 10:54:10.296620 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" event={"ID":"03debb58-d38a-4470-9274-57f7a61f622f","Type":"ContainerDied","Data":"5f399b212cbb7a69f6e81ec179777bdf6f8e0d3496393cb1ba1a8bd8e02b6345"} Feb 28 10:54:11 crc kubenswrapper[4972]: I0228 10:54:11.857157 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.544433 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.544928 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7bh5cch5d4h66ch66dh57bh669h6h699h7dh665h64bh59fh55fh67h5d7hc4h689h544h74h5b4h595h79h67fh57dh549h5c7h5bbh66dh555h587h58q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxrzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5db7884c8f-qvwtf_openstack(96039bc0-559a-464e-bc9c-f79fba169a2b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.565923 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5db7884c8f-qvwtf" podUID="96039bc0-559a-464e-bc9c-f79fba169a2b" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.640715 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.640906 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ncdh574h54hd9h5h5dchb6h586h64bh58dh578h664h56dh5ffh5b8h7dh8chbch64ch674h66fh647h668h68chc9h5b7h7dh59dhc9hbch59fh584q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x7f7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-54f8c57c9c-kq5wx_openstack(ae5e2238-8b10-439d-a815-610ee23d23fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.645293 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-54f8c57c9c-kq5wx" podUID="ae5e2238-8b10-439d-a815-610ee23d23fc" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.655827 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.656002 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dfh56bh648h67ch687h5ddh55dhd5h5d9h656h6ch668hdh6bhb5h59ch557h576h69h5fdhf4h677hdh59fh596h5b5h648h96h676h5f4hc8hbfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jcsdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6d76fdd467-knmck_openstack(f20522e2-5d4f-45d9-92e0-c51892b78a41): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:16 crc kubenswrapper[4972]: E0228 10:54:16.659537 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6d76fdd467-knmck" podUID="f20522e2-5d4f-45d9-92e0-c51892b78a41" Feb 28 10:54:16 crc kubenswrapper[4972]: I0228 10:54:16.683979 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:16 crc kubenswrapper[4972]: I0228 10:54:16.852139 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcbgb\" (UniqueName: \"kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb\") pod \"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3\" (UID: \"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3\") " Feb 28 10:54:16 crc kubenswrapper[4972]: I0228 10:54:16.857303 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Feb 28 10:54:16 crc kubenswrapper[4972]: I0228 10:54:16.870143 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb" (OuterVolumeSpecName: "kube-api-access-xcbgb") pod "625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" (UID: "625f85df-e9dd-4c46-9a05-b4da0c0bd5a3"). InnerVolumeSpecName "kube-api-access-xcbgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:16 crc kubenswrapper[4972]: I0228 10:54:16.956621 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcbgb\" (UniqueName: \"kubernetes.io/projected/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3-kube-api-access-xcbgb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:17 crc kubenswrapper[4972]: I0228 10:54:17.365348 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" Feb 28 10:54:17 crc kubenswrapper[4972]: I0228 10:54:17.365394 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537934-2kjkh" event={"ID":"625f85df-e9dd-4c46-9a05-b4da0c0bd5a3","Type":"ContainerDied","Data":"bc167efc253c6fcbe0b41df7f2cb4e387c011b8f111025655a9529e53c9d9059"} Feb 28 10:54:17 crc kubenswrapper[4972]: I0228 10:54:17.365423 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc167efc253c6fcbe0b41df7f2cb4e387c011b8f111025655a9529e53c9d9059" Feb 28 10:54:17 crc kubenswrapper[4972]: I0228 10:54:17.776715 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537928-lpsh8"] Feb 28 10:54:17 crc kubenswrapper[4972]: I0228 10:54:17.807949 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537928-lpsh8"] Feb 28 10:54:19 crc kubenswrapper[4972]: I0228 10:54:19.802181 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96ffb7ef-d239-4a03-8654-641bcad251bd" path="/var/lib/kubelet/pods/96ffb7ef-d239-4a03-8654-641bcad251bd/volumes" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.054840 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164218 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164321 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164379 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164452 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntdxq\" (UniqueName: \"kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164599 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.164846 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data\") pod \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\" (UID: \"53609995-2a99-40b6-8ef7-bbeb7f20eb89\") " Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.173894 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts" (OuterVolumeSpecName: "scripts") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.173945 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.174046 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.184762 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq" (OuterVolumeSpecName: "kube-api-access-ntdxq") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "kube-api-access-ntdxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.193030 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data" (OuterVolumeSpecName: "config-data") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.194154 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53609995-2a99-40b6-8ef7-bbeb7f20eb89" (UID: "53609995-2a99-40b6-8ef7-bbeb7f20eb89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268197 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268237 4972 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268247 4972 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268259 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268269 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntdxq\" (UniqueName: \"kubernetes.io/projected/53609995-2a99-40b6-8ef7-bbeb7f20eb89-kube-api-access-ntdxq\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.268280 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53609995-2a99-40b6-8ef7-bbeb7f20eb89-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.404189 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vkvhp" event={"ID":"53609995-2a99-40b6-8ef7-bbeb7f20eb89","Type":"ContainerDied","Data":"e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1"} Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.404234 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9b8a79356b8572aa17bce1b56b69007664c132d2739141c7385e5a43fc0d2f1" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.404339 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vkvhp" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.857088 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Feb 28 10:54:21 crc kubenswrapper[4972]: I0228 10:54:21.857247 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.136121 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vkvhp"] Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.145715 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vkvhp"] Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.247250 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wwvfq"] Feb 28 10:54:22 crc kubenswrapper[4972]: E0228 10:54:22.247965 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53609995-2a99-40b6-8ef7-bbeb7f20eb89" containerName="keystone-bootstrap" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.247989 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="53609995-2a99-40b6-8ef7-bbeb7f20eb89" containerName="keystone-bootstrap" Feb 28 10:54:22 crc kubenswrapper[4972]: E0228 10:54:22.248014 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" containerName="oc" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.248023 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" containerName="oc" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.248238 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" containerName="oc" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.248264 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="53609995-2a99-40b6-8ef7-bbeb7f20eb89" containerName="keystone-bootstrap" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.249017 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.253245 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.253763 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.253894 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.257163 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.257737 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9794v" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.260439 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wwvfq"] Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.399258 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.399448 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lvzp\" (UniqueName: \"kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.399665 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.399795 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.399991 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.400019 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502316 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502443 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lvzp\" (UniqueName: \"kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502515 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502541 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.502590 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.508166 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.508480 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.509478 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.512919 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.518972 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.520721 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lvzp\" (UniqueName: \"kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp\") pod \"keystone-bootstrap-wwvfq\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:22 crc kubenswrapper[4972]: I0228 10:54:22.573759 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:23 crc kubenswrapper[4972]: I0228 10:54:23.802675 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53609995-2a99-40b6-8ef7-bbeb7f20eb89" path="/var/lib/kubelet/pods/53609995-2a99-40b6-8ef7-bbeb7f20eb89/volumes" Feb 28 10:54:26 crc kubenswrapper[4972]: I0228 10:54:26.856664 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Feb 28 10:54:28 crc kubenswrapper[4972]: I0228 10:54:28.890276 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:54:28 crc kubenswrapper[4972]: I0228 10:54:28.890706 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.025866 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.163044 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7f7c\" (UniqueName: \"kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c\") pod \"ae5e2238-8b10-439d-a815-610ee23d23fc\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.163157 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts\") pod \"ae5e2238-8b10-439d-a815-610ee23d23fc\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.163261 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs\") pod \"ae5e2238-8b10-439d-a815-610ee23d23fc\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.163304 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key\") pod \"ae5e2238-8b10-439d-a815-610ee23d23fc\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.163382 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data\") pod \"ae5e2238-8b10-439d-a815-610ee23d23fc\" (UID: \"ae5e2238-8b10-439d-a815-610ee23d23fc\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.164275 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data" (OuterVolumeSpecName: "config-data") pod "ae5e2238-8b10-439d-a815-610ee23d23fc" (UID: "ae5e2238-8b10-439d-a815-610ee23d23fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.173145 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs" (OuterVolumeSpecName: "logs") pod "ae5e2238-8b10-439d-a815-610ee23d23fc" (UID: "ae5e2238-8b10-439d-a815-610ee23d23fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.173237 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts" (OuterVolumeSpecName: "scripts") pod "ae5e2238-8b10-439d-a815-610ee23d23fc" (UID: "ae5e2238-8b10-439d-a815-610ee23d23fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.176059 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ae5e2238-8b10-439d-a815-610ee23d23fc" (UID: "ae5e2238-8b10-439d-a815-610ee23d23fc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.176607 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c" (OuterVolumeSpecName: "kube-api-access-x7f7c") pod "ae5e2238-8b10-439d-a815-610ee23d23fc" (UID: "ae5e2238-8b10-439d-a815-610ee23d23fc"). InnerVolumeSpecName "kube-api-access-x7f7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.266042 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae5e2238-8b10-439d-a815-610ee23d23fc-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.266433 4972 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae5e2238-8b10-439d-a815-610ee23d23fc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.266447 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.266471 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7f7c\" (UniqueName: \"kubernetes.io/projected/ae5e2238-8b10-439d-a815-610ee23d23fc-kube-api-access-x7f7c\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.266484 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae5e2238-8b10-439d-a815-610ee23d23fc-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: E0228 10:54:29.365995 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 28 10:54:29 crc kubenswrapper[4972]: E0228 10:54:29.366219 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67h64ch584h5d6h8fh574hbbh545h5f4h5c6h665hdh9ch554h664h6h5h58ch5bfh674h667h64fh586hb7h675h68bh86h7chddh89h5h55bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t5tlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(84e043f8-8906-4baf-a610-53446e9c0f85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.371825 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.384947 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.389621 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.400813 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471314 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs\") pod \"96039bc0-559a-464e-bc9c-f79fba169a2b\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471380 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471480 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471500 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471526 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxrzk\" (UniqueName: \"kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk\") pod \"96039bc0-559a-464e-bc9c-f79fba169a2b\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471563 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471603 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471628 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data\") pod \"96039bc0-559a-464e-bc9c-f79fba169a2b\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471696 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58d29\" (UniqueName: \"kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471755 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts\") pod \"96039bc0-559a-464e-bc9c-f79fba169a2b\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471775 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471791 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs\") pod \"b92a57f4-00aa-404f-8263-b0c8046c6811\" (UID: \"b92a57f4-00aa-404f-8263-b0c8046c6811\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.471816 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key\") pod \"96039bc0-559a-464e-bc9c-f79fba169a2b\" (UID: \"96039bc0-559a-464e-bc9c-f79fba169a2b\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.472305 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs" (OuterVolumeSpecName: "logs") pod "96039bc0-559a-464e-bc9c-f79fba169a2b" (UID: "96039bc0-559a-464e-bc9c-f79fba169a2b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.472377 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs" (OuterVolumeSpecName: "logs") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.472801 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts" (OuterVolumeSpecName: "scripts") pod "96039bc0-559a-464e-bc9c-f79fba169a2b" (UID: "96039bc0-559a-464e-bc9c-f79fba169a2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.473026 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.473384 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data" (OuterVolumeSpecName: "config-data") pod "96039bc0-559a-464e-bc9c-f79fba169a2b" (UID: "96039bc0-559a-464e-bc9c-f79fba169a2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.477104 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "96039bc0-559a-464e-bc9c-f79fba169a2b" (UID: "96039bc0-559a-464e-bc9c-f79fba169a2b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.477204 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.477245 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29" (OuterVolumeSpecName: "kube-api-access-58d29") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "kube-api-access-58d29". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.482671 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts" (OuterVolumeSpecName: "scripts") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.485414 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk" (OuterVolumeSpecName: "kube-api-access-wxrzk") pod "96039bc0-559a-464e-bc9c-f79fba169a2b" (UID: "96039bc0-559a-464e-bc9c-f79fba169a2b"). InnerVolumeSpecName "kube-api-access-wxrzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.492659 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5db7884c8f-qvwtf" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.493404 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5db7884c8f-qvwtf" event={"ID":"96039bc0-559a-464e-bc9c-f79fba169a2b","Type":"ContainerDied","Data":"2e0da091dcf8dbbb9d5d1926ea645aa94c5c997d8c6a15e1af45e11667ce6200"} Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.505275 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.505413 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92a57f4-00aa-404f-8263-b0c8046c6811","Type":"ContainerDied","Data":"1924fba11ac268c737dc798aebe984ec9768bebefbd4d0cde9bb1b18dd350b97"} Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.505501 4972 scope.go:117] "RemoveContainer" containerID="c6348203524815c0eaf13d9c19ebc28ef73b33576c027466ab98c35bc0fd53c2" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.505452 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.510867 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"357785c0-3dd9-4364-9f53-89bde0b2d7d2","Type":"ContainerDied","Data":"c26dec33e35e74b5f7b178f48c332501c41be9988288e21f1009590a15e17b68"} Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.510916 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.514943 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d76fdd467-knmck" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.514950 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d76fdd467-knmck" event={"ID":"f20522e2-5d4f-45d9-92e0-c51892b78a41","Type":"ContainerDied","Data":"4450cf1348dfd8006a6f71e720a82c2631f06f410c23627a8ab84a3a07c19c2b"} Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.520002 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54f8c57c9c-kq5wx" event={"ID":"ae5e2238-8b10-439d-a815-610ee23d23fc","Type":"ContainerDied","Data":"e2c04469f9d5e8b08b68dc56839f5b7cc141867d48a7aa793b91a8e78e0ada3a"} Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.520057 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54f8c57c9c-kq5wx" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.540772 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data" (OuterVolumeSpecName: "config-data") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.549115 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b92a57f4-00aa-404f-8263-b0c8046c6811" (UID: "b92a57f4-00aa-404f-8263-b0c8046c6811"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.572837 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key\") pod \"f20522e2-5d4f-45d9-92e0-c51892b78a41\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.572915 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573002 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573050 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs\") pod \"f20522e2-5d4f-45d9-92e0-c51892b78a41\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573104 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573142 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573191 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcsdq\" (UniqueName: \"kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq\") pod \"f20522e2-5d4f-45d9-92e0-c51892b78a41\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573254 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skgnf\" (UniqueName: \"kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573314 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573348 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573394 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts\") pod \"f20522e2-5d4f-45d9-92e0-c51892b78a41\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573446 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data\") pod \"f20522e2-5d4f-45d9-92e0-c51892b78a41\" (UID: \"f20522e2-5d4f-45d9-92e0-c51892b78a41\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.573497 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run\") pod \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\" (UID: \"357785c0-3dd9-4364-9f53-89bde0b2d7d2\") " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574558 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs" (OuterVolumeSpecName: "logs") pod "f20522e2-5d4f-45d9-92e0-c51892b78a41" (UID: "f20522e2-5d4f-45d9-92e0-c51892b78a41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574904 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58d29\" (UniqueName: \"kubernetes.io/projected/b92a57f4-00aa-404f-8263-b0c8046c6811-kube-api-access-58d29\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574928 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574942 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574954 4972 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574966 4972 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/96039bc0-559a-464e-bc9c-f79fba169a2b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574978 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96039bc0-559a-464e-bc9c-f79fba169a2b-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.574988 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575000 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575080 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575095 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxrzk\" (UniqueName: \"kubernetes.io/projected/96039bc0-559a-464e-bc9c-f79fba169a2b-kube-api-access-wxrzk\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575108 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92a57f4-00aa-404f-8263-b0c8046c6811-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575123 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92a57f4-00aa-404f-8263-b0c8046c6811-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.575135 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96039bc0-559a-464e-bc9c-f79fba169a2b-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.577174 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f20522e2-5d4f-45d9-92e0-c51892b78a41" (UID: "f20522e2-5d4f-45d9-92e0-c51892b78a41"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.577865 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs" (OuterVolumeSpecName: "logs") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.578486 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.579057 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.579248 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts" (OuterVolumeSpecName: "scripts") pod "f20522e2-5d4f-45d9-92e0-c51892b78a41" (UID: "f20522e2-5d4f-45d9-92e0-c51892b78a41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.579523 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data" (OuterVolumeSpecName: "config-data") pod "f20522e2-5d4f-45d9-92e0-c51892b78a41" (UID: "f20522e2-5d4f-45d9-92e0-c51892b78a41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.584003 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts" (OuterVolumeSpecName: "scripts") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.584078 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf" (OuterVolumeSpecName: "kube-api-access-skgnf") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "kube-api-access-skgnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.586226 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.594753 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq" (OuterVolumeSpecName: "kube-api-access-jcsdq") pod "f20522e2-5d4f-45d9-92e0-c51892b78a41" (UID: "f20522e2-5d4f-45d9-92e0-c51892b78a41"). InnerVolumeSpecName "kube-api-access-jcsdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.611049 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5db7884c8f-qvwtf"] Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.630798 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.631175 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.636281 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.639219 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-54f8c57c9c-kq5wx"] Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.639311 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data" (OuterVolumeSpecName: "config-data") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677183 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677219 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677228 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f20522e2-5d4f-45d9-92e0-c51892b78a41-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677237 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677246 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677255 4972 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f20522e2-5d4f-45d9-92e0-c51892b78a41-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677288 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677301 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f20522e2-5d4f-45d9-92e0-c51892b78a41-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677313 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/357785c0-3dd9-4364-9f53-89bde0b2d7d2-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677324 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677336 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcsdq\" (UniqueName: \"kubernetes.io/projected/f20522e2-5d4f-45d9-92e0-c51892b78a41-kube-api-access-jcsdq\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677347 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skgnf\" (UniqueName: \"kubernetes.io/projected/357785c0-3dd9-4364-9f53-89bde0b2d7d2-kube-api-access-skgnf\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.677358 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.678704 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "357785c0-3dd9-4364-9f53-89bde0b2d7d2" (UID: "357785c0-3dd9-4364-9f53-89bde0b2d7d2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.696542 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.782565 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.782609 4972 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/357785c0-3dd9-4364-9f53-89bde0b2d7d2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.801438 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96039bc0-559a-464e-bc9c-f79fba169a2b" path="/var/lib/kubelet/pods/96039bc0-559a-464e-bc9c-f79fba169a2b/volumes" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.801913 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae5e2238-8b10-439d-a815-610ee23d23fc" path="/var/lib/kubelet/pods/ae5e2238-8b10-439d-a815-610ee23d23fc/volumes" Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.913446 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:29 crc kubenswrapper[4972]: I0228 10:54:29.931633 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d76fdd467-knmck"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.021208 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.056593 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.068799 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.080243 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.090294 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: E0228 10:54:30.090801 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.090823 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: E0228 10:54:30.090843 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.090852 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: E0228 10:54:30.090866 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.090872 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: E0228 10:54:30.090885 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.090891 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.091068 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.091088 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-log" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.091102 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.091118 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" containerName="glance-httpd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.092444 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.094628 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.096746 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gsngd" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.096751 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.097022 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.102687 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.105656 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.109241 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.109417 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.117973 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.144325 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202402 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202476 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202525 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202597 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202647 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqvpv\" (UniqueName: \"kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202702 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.202792 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203097 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203176 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203262 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203378 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203432 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203508 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203589 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxtm4\" (UniqueName: \"kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203655 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.203680 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.305233 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxtm4\" (UniqueName: \"kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.305300 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306244 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306271 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306304 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306330 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306361 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306376 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqvpv\" (UniqueName: \"kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306392 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306485 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306509 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306543 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306582 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306663 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306683 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.306712 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.307839 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.314950 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.327212 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.327414 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.328598 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.331184 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.331398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.331559 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.334659 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.335678 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.336577 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.338104 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqvpv\" (UniqueName: \"kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.340288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxtm4\" (UniqueName: \"kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.346332 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.353603 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.354911 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.359419 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " pod="openstack/glance-default-external-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.366388 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.424015 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:30 crc kubenswrapper[4972]: I0228 10:54:30.428294 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:54:31 crc kubenswrapper[4972]: I0228 10:54:31.802316 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="357785c0-3dd9-4364-9f53-89bde0b2d7d2" path="/var/lib/kubelet/pods/357785c0-3dd9-4364-9f53-89bde0b2d7d2/volumes" Feb 28 10:54:31 crc kubenswrapper[4972]: I0228 10:54:31.803754 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92a57f4-00aa-404f-8263-b0c8046c6811" path="/var/lib/kubelet/pods/b92a57f4-00aa-404f-8263-b0c8046c6811/volumes" Feb 28 10:54:31 crc kubenswrapper[4972]: I0228 10:54:31.804796 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f20522e2-5d4f-45d9-92e0-c51892b78a41" path="/var/lib/kubelet/pods/f20522e2-5d4f-45d9-92e0-c51892b78a41/volumes" Feb 28 10:54:32 crc kubenswrapper[4972]: E0228 10:54:32.547358 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 28 10:54:32 crc kubenswrapper[4972]: E0228 10:54:32.548227 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zrtws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-qtm2c_openstack(aa1673c6-2e2d-496f-b152-ca9df91df6d7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:32 crc kubenswrapper[4972]: E0228 10:54:32.549663 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-qtm2c" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" Feb 28 10:54:33 crc kubenswrapper[4972]: E0228 10:54:33.569288 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-qtm2c" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" Feb 28 10:54:33 crc kubenswrapper[4972]: E0228 10:54:33.750397 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 28 10:54:33 crc kubenswrapper[4972]: E0228 10:54:33.750604 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25ds5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-llqn9_openstack(edd35ce5-02ce-4909-89c5-6dfce27e50a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 10:54:33 crc kubenswrapper[4972]: E0228 10:54:33.751977 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-llqn9" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" Feb 28 10:54:33 crc kubenswrapper[4972]: I0228 10:54:33.790168 4972 scope.go:117] "RemoveContainer" containerID="e2a3ff4c70fff3c0486ff02e44988c3ef586bd11d5539f999aa6aabf02cc0f3a" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.050884 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.052098 4972 scope.go:117] "RemoveContainer" containerID="4ba5d0491a86ef873e6bf6b3c66dd4e5e40b87fe90a545af21ee4343e786469c" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.096772 4972 scope.go:117] "RemoveContainer" containerID="2bca29557d677209fa721dbc218c9fc183d20bd48f31eddc244b8dcddc8809f2" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.226629 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.227259 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.227374 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bllmg\" (UniqueName: \"kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.227412 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.227453 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.227567 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb\") pod \"03debb58-d38a-4470-9274-57f7a61f622f\" (UID: \"03debb58-d38a-4470-9274-57f7a61f622f\") " Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.244839 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg" (OuterVolumeSpecName: "kube-api-access-bllmg") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "kube-api-access-bllmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.277982 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.286848 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.291580 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.294677 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.307478 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config" (OuterVolumeSpecName: "config") pod "03debb58-d38a-4470-9274-57f7a61f622f" (UID: "03debb58-d38a-4470-9274-57f7a61f622f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331057 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331358 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331438 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331513 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bllmg\" (UniqueName: \"kubernetes.io/projected/03debb58-d38a-4470-9274-57f7a61f622f-kube-api-access-bllmg\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331567 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.331628 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03debb58-d38a-4470-9274-57f7a61f622f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.337545 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wwvfq"] Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.355781 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c6cf8f456-722rs"] Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.365978 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.583263 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" event={"ID":"03debb58-d38a-4470-9274-57f7a61f622f","Type":"ContainerDied","Data":"353609d79ce2681e435d7ea592d1f92e0b28f53ea01c1db3082a7a37b5b033c6"} Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.583281 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.583345 4972 scope.go:117] "RemoveContainer" containerID="5f399b212cbb7a69f6e81ec179777bdf6f8e0d3496393cb1ba1a8bd8e02b6345" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.588552 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kmth" event={"ID":"7dc233b8-4780-4e84-a0b6-dd5fe2289e16","Type":"ContainerStarted","Data":"d698e3cd64e4a29016b5708ec7fd450c9907acdd34c11aa14d442451c324ed0a"} Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.595513 4972 generic.go:334] "Generic (PLEG): container finished" podID="ef9e4339-b034-40f0-a1f0-04229463c846" containerID="39215e36b6acf94dfdcb4712e55fdad5b1116ed4b55c3f2229e2c570a7079184" exitCode=0 Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.595619 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vppxs" event={"ID":"ef9e4339-b034-40f0-a1f0-04229463c846","Type":"ContainerDied","Data":"39215e36b6acf94dfdcb4712e55fdad5b1116ed4b55c3f2229e2c570a7079184"} Feb 28 10:54:34 crc kubenswrapper[4972]: E0228 10:54:34.602060 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-llqn9" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.625651 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.627334 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4kmth" podStartSLOduration=7.230013489 podStartE2EDuration="35.627321525s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="2026-02-28 10:54:00.9571653 +0000 UTC m=+1097.869173048" lastFinishedPulling="2026-02-28 10:54:29.354473346 +0000 UTC m=+1126.266481084" observedRunningTime="2026-02-28 10:54:34.612178713 +0000 UTC m=+1131.524186451" watchObservedRunningTime="2026-02-28 10:54:34.627321525 +0000 UTC m=+1131.539329263" Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.640347 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:54:34 crc kubenswrapper[4972]: W0228 10:54:34.641737 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ce78cca_b508_4c23_b331_21f1840ba47c.slice/crio-e5000cce091bf3e88641440b6d8f6e4032745a604c00f297473e69cdbcf3dbcb WatchSource:0}: Error finding container e5000cce091bf3e88641440b6d8f6e4032745a604c00f297473e69cdbcf3dbcb: Status 404 returned error can't find the container with id e5000cce091bf3e88641440b6d8f6e4032745a604c00f297473e69cdbcf3dbcb Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.652581 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tgx4k"] Feb 28 10:54:34 crc kubenswrapper[4972]: I0228 10:54:34.672877 4972 scope.go:117] "RemoveContainer" containerID="d153b4cc65055ec3c48638b1ee3c7c1cce21541dabe147c68826e29bbde2a016" Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.620361 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerStarted","Data":"e5000cce091bf3e88641440b6d8f6e4032745a604c00f297473e69cdbcf3dbcb"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.625695 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerStarted","Data":"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.625938 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerStarted","Data":"e5ce21101515d7be8f05d51d461295627f4f68e109e8ea06b4e7503a2e605591"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.636983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6cf8f456-722rs" event={"ID":"a24fd1fd-763d-4106-87b0-805e2f1fc6c5","Type":"ContainerStarted","Data":"f150b1ed887226f0178558ca43255cae05a3d932122fc24f7da51c420af9fda2"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.637054 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6cf8f456-722rs" event={"ID":"a24fd1fd-763d-4106-87b0-805e2f1fc6c5","Type":"ContainerStarted","Data":"927b73257d75e24875ca29dee39ae79e44e879756670041cb8081ff4ed6fed8c"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.637070 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6cf8f456-722rs" event={"ID":"a24fd1fd-763d-4106-87b0-805e2f1fc6c5","Type":"ContainerStarted","Data":"cb5a4a5e7e8b25d7a7dccd6fbf69b7a731b46ef2a6180be666603ac0755928c5"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.639827 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerStarted","Data":"9b5761dfc835d28b55d4bd522cef3703451247ab368d78fd87ea8706db0cec97"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.647058 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wwvfq" event={"ID":"4d80893f-182b-4755-afea-fd6635c700fd","Type":"ContainerStarted","Data":"e397650d7d4ff9605e399e9aa184c180ebf09875d4634e14ffedd73ac5d41e04"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.647105 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wwvfq" event={"ID":"4d80893f-182b-4755-afea-fd6635c700fd","Type":"ContainerStarted","Data":"ce9560f73438f96621ddeeb6b2ba93ec4e1363bae4c22aca21f064785d4ac0b9"} Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.676704 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.689757 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c6cf8f456-722rs" podStartSLOduration=27.240802154 podStartE2EDuration="27.68972768s" podCreationTimestamp="2026-02-28 10:54:08 +0000 UTC" firstStartedPulling="2026-02-28 10:54:34.642608241 +0000 UTC m=+1131.554615979" lastFinishedPulling="2026-02-28 10:54:35.091533767 +0000 UTC m=+1132.003541505" observedRunningTime="2026-02-28 10:54:35.664727255 +0000 UTC m=+1132.576734993" watchObservedRunningTime="2026-02-28 10:54:35.68972768 +0000 UTC m=+1132.601735418" Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.694645 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wwvfq" podStartSLOduration=13.694620517 podStartE2EDuration="13.694620517s" podCreationTimestamp="2026-02-28 10:54:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:35.692161328 +0000 UTC m=+1132.604169066" watchObservedRunningTime="2026-02-28 10:54:35.694620517 +0000 UTC m=+1132.606628255" Feb 28 10:54:35 crc kubenswrapper[4972]: W0228 10:54:35.707672 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb0b990a_60c3_4d43_9f11_883155d37a5a.slice/crio-bf02cc5f0126e963966b788c404a617d431712af189fad6061b19640b6c99db8 WatchSource:0}: Error finding container bf02cc5f0126e963966b788c404a617d431712af189fad6061b19640b6c99db8: Status 404 returned error can't find the container with id bf02cc5f0126e963966b788c404a617d431712af189fad6061b19640b6c99db8 Feb 28 10:54:35 crc kubenswrapper[4972]: I0228 10:54:35.828784 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03debb58-d38a-4470-9274-57f7a61f622f" path="/var/lib/kubelet/pods/03debb58-d38a-4470-9274-57f7a61f622f/volumes" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.140450 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vppxs" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.279656 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config\") pod \"ef9e4339-b034-40f0-a1f0-04229463c846\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.279753 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdsmf\" (UniqueName: \"kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf\") pod \"ef9e4339-b034-40f0-a1f0-04229463c846\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.279822 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle\") pod \"ef9e4339-b034-40f0-a1f0-04229463c846\" (UID: \"ef9e4339-b034-40f0-a1f0-04229463c846\") " Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.291933 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf" (OuterVolumeSpecName: "kube-api-access-sdsmf") pod "ef9e4339-b034-40f0-a1f0-04229463c846" (UID: "ef9e4339-b034-40f0-a1f0-04229463c846"). InnerVolumeSpecName "kube-api-access-sdsmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.323587 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef9e4339-b034-40f0-a1f0-04229463c846" (UID: "ef9e4339-b034-40f0-a1f0-04229463c846"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.328793 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config" (OuterVolumeSpecName: "config") pod "ef9e4339-b034-40f0-a1f0-04229463c846" (UID: "ef9e4339-b034-40f0-a1f0-04229463c846"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.382111 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.382164 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdsmf\" (UniqueName: \"kubernetes.io/projected/ef9e4339-b034-40f0-a1f0-04229463c846-kube-api-access-sdsmf\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.382180 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9e4339-b034-40f0-a1f0-04229463c846-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.664374 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerStarted","Data":"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.673558 4972 generic.go:334] "Generic (PLEG): container finished" podID="7dc233b8-4780-4e84-a0b6-dd5fe2289e16" containerID="d698e3cd64e4a29016b5708ec7fd450c9907acdd34c11aa14d442451c324ed0a" exitCode=0 Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.673611 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kmth" event={"ID":"7dc233b8-4780-4e84-a0b6-dd5fe2289e16","Type":"ContainerDied","Data":"d698e3cd64e4a29016b5708ec7fd450c9907acdd34c11aa14d442451c324ed0a"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.676298 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vppxs" event={"ID":"ef9e4339-b034-40f0-a1f0-04229463c846","Type":"ContainerDied","Data":"924c6704501cbaf038d4b6be13d2b23aa1d4f5084955fe278590b962e5ecd346"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.676362 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="924c6704501cbaf038d4b6be13d2b23aa1d4f5084955fe278590b962e5ecd346" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.676477 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vppxs" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.702183 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerStarted","Data":"5e1a8e0c6c8ad4ffa9e50e451cfe899f0a2f0f316295b1f5df98c6b1089d1226"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.702241 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerStarted","Data":"ed4586be8bb58a2b22c320a60174e676c9f5a97c0ae37f6505a39f16bad92ba7"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.703590 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c898cfcfb-6nt9v" podStartSLOduration=29.170720839 podStartE2EDuration="29.703563773s" podCreationTimestamp="2026-02-28 10:54:07 +0000 UTC" firstStartedPulling="2026-02-28 10:54:34.64116008 +0000 UTC m=+1131.553167818" lastFinishedPulling="2026-02-28 10:54:35.174003014 +0000 UTC m=+1132.086010752" observedRunningTime="2026-02-28 10:54:36.695392776 +0000 UTC m=+1133.607400514" watchObservedRunningTime="2026-02-28 10:54:36.703563773 +0000 UTC m=+1133.615571511" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.714550 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerStarted","Data":"1c72d419d0c725671b6225463290a64838849df9bfd96d7ac9b36684ca917807"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.714594 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerStarted","Data":"bf02cc5f0126e963966b788c404a617d431712af189fad6061b19640b6c99db8"} Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.761198 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.7611773490000004 podStartE2EDuration="7.761177349s" podCreationTimestamp="2026-02-28 10:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:36.757479906 +0000 UTC m=+1133.669487664" watchObservedRunningTime="2026-02-28 10:54:36.761177349 +0000 UTC m=+1133.673185087" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.857094 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-tgx4k" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.916533 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:54:36 crc kubenswrapper[4972]: E0228 10:54:36.917048 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9e4339-b034-40f0-a1f0-04229463c846" containerName="neutron-db-sync" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.917064 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9e4339-b034-40f0-a1f0-04229463c846" containerName="neutron-db-sync" Feb 28 10:54:36 crc kubenswrapper[4972]: E0228 10:54:36.917076 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.917081 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" Feb 28 10:54:36 crc kubenswrapper[4972]: E0228 10:54:36.917107 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="init" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.917118 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="init" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.917290 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9e4339-b034-40f0-a1f0-04229463c846" containerName="neutron-db-sync" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.917319 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="03debb58-d38a-4470-9274-57f7a61f622f" containerName="dnsmasq-dns" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.918350 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:36 crc kubenswrapper[4972]: I0228 10:54:36.944295 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002321 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002442 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002490 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002552 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002588 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjls5\" (UniqueName: \"kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.002660 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.006969 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.010312 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.017685 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.017886 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.018023 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.018138 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jdxf6" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.038999 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.109930 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjls5\" (UniqueName: \"kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110000 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110032 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110083 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110116 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110170 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110254 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110289 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110333 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tw5w\" (UniqueName: \"kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.110362 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.111221 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.112180 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.112749 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.113248 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.113770 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.150815 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjls5\" (UniqueName: \"kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5\") pod \"dnsmasq-dns-55f844cf75-b5sjv\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.214605 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.214667 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tw5w\" (UniqueName: \"kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.214702 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.214719 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.214754 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.231436 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.232289 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.241135 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.241818 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.274001 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.274285 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tw5w\" (UniqueName: \"kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w\") pod \"neutron-564fbcbf5b-lsdwn\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.353268 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.745735 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerStarted","Data":"ad5e9d261027f103027150480385caf8114daa4369d7c7c2646d66eaee8a2f94"} Feb 28 10:54:37 crc kubenswrapper[4972]: I0228 10:54:37.895600 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.89557242 podStartE2EDuration="8.89557242s" podCreationTimestamp="2026-02-28 10:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:37.866299764 +0000 UTC m=+1134.778307502" watchObservedRunningTime="2026-02-28 10:54:37.89557242 +0000 UTC m=+1134.807580168" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.025899 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.236903 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.321627 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kmth" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.376497 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.376600 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.385056 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts\") pod \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.385151 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs\") pod \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.385250 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data\") pod \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.385885 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zzxx\" (UniqueName: \"kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx\") pod \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.386148 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle\") pod \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\" (UID: \"7dc233b8-4780-4e84-a0b6-dd5fe2289e16\") " Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.387931 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs" (OuterVolumeSpecName: "logs") pod "7dc233b8-4780-4e84-a0b6-dd5fe2289e16" (UID: "7dc233b8-4780-4e84-a0b6-dd5fe2289e16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.396808 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx" (OuterVolumeSpecName: "kube-api-access-2zzxx") pod "7dc233b8-4780-4e84-a0b6-dd5fe2289e16" (UID: "7dc233b8-4780-4e84-a0b6-dd5fe2289e16"). InnerVolumeSpecName "kube-api-access-2zzxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.403950 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts" (OuterVolumeSpecName: "scripts") pod "7dc233b8-4780-4e84-a0b6-dd5fe2289e16" (UID: "7dc233b8-4780-4e84-a0b6-dd5fe2289e16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.434359 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dc233b8-4780-4e84-a0b6-dd5fe2289e16" (UID: "7dc233b8-4780-4e84-a0b6-dd5fe2289e16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.488523 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zzxx\" (UniqueName: \"kubernetes.io/projected/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-kube-api-access-2zzxx\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.488566 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.488575 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.488584 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.501582 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data" (OuterVolumeSpecName: "config-data") pod "7dc233b8-4780-4e84-a0b6-dd5fe2289e16" (UID: "7dc233b8-4780-4e84-a0b6-dd5fe2289e16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.554997 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.555659 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.590677 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dc233b8-4780-4e84-a0b6-dd5fe2289e16-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.838630 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerID="bf116311e059a34e36dcb51d90e94848584b74ec778f7dc7aa3afcc4af7116f1" exitCode=0 Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.838703 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" event={"ID":"0bc05540-b9af-4d90-abd7-2b6d4d8ef976","Type":"ContainerDied","Data":"bf116311e059a34e36dcb51d90e94848584b74ec778f7dc7aa3afcc4af7116f1"} Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.838732 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" event={"ID":"0bc05540-b9af-4d90-abd7-2b6d4d8ef976","Type":"ContainerStarted","Data":"fee32373a4e5c66360bc1856106c63c7b5e14c9ae94bdf4846f53385c32849ae"} Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.857438 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerStarted","Data":"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701"} Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.857518 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerStarted","Data":"4f89459f4579cc3177b12241f938868587e3538a6c0b005bb0c8f66f8dfce383"} Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.872158 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kmth" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.878208 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kmth" event={"ID":"7dc233b8-4780-4e84-a0b6-dd5fe2289e16","Type":"ContainerDied","Data":"c86efac22b4a5c2c031a278c8f9a6ad8190d167c7444dd2a470fbf5f85adfe34"} Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.878264 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c86efac22b4a5c2c031a278c8f9a6ad8190d167c7444dd2a470fbf5f85adfe34" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.889584 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:54:38 crc kubenswrapper[4972]: E0228 10:54:38.890195 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc233b8-4780-4e84-a0b6-dd5fe2289e16" containerName="placement-db-sync" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.890218 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc233b8-4780-4e84-a0b6-dd5fe2289e16" containerName="placement-db-sync" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.895098 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc233b8-4780-4e84-a0b6-dd5fe2289e16" containerName="placement-db-sync" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.896603 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.912512 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.915264 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.915347 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.922044 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.927102 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wp8gv" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.928675 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.998593 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999099 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmn6j\" (UniqueName: \"kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999125 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999186 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999302 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999368 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:38 crc kubenswrapper[4972]: I0228 10:54:38.999405 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102614 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102683 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmn6j\" (UniqueName: \"kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102706 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102732 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102824 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102872 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.102899 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.105037 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.131938 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.140910 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.146272 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmn6j\" (UniqueName: \"kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.146994 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.147183 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.147693 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data\") pod \"placement-5cf557df54-q4jtg\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.344962 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.895657 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.906542 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerStarted","Data":"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9"} Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.906602 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.931985 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" event={"ID":"0bc05540-b9af-4d90-abd7-2b6d4d8ef976","Type":"ContainerStarted","Data":"202f19364f0f0b5b5babf7a2d417cfe899433378faf2c7d685894c7929a8cce0"} Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.932032 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.948999 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-564fbcbf5b-lsdwn" podStartSLOduration=3.948979753 podStartE2EDuration="3.948979753s" podCreationTimestamp="2026-02-28 10:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:39.947421149 +0000 UTC m=+1136.859428887" watchObservedRunningTime="2026-02-28 10:54:39.948979753 +0000 UTC m=+1136.860987501" Feb 28 10:54:39 crc kubenswrapper[4972]: I0228 10:54:39.997829 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" podStartSLOduration=3.997802493 podStartE2EDuration="3.997802493s" podCreationTimestamp="2026-02-28 10:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:39.988382281 +0000 UTC m=+1136.900390029" watchObservedRunningTime="2026-02-28 10:54:39.997802493 +0000 UTC m=+1136.909810241" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.424193 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.424630 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.429408 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.429439 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.474266 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.478854 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.491664 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.516762 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.763201 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.765599 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.769430 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.772613 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.775952 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.870887 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871345 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871398 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871487 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871506 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871524 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.871540 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68w4b\" (UniqueName: \"kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976119 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976224 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976287 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976358 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976380 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976431 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.976476 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68w4b\" (UniqueName: \"kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:40 crc kubenswrapper[4972]: I0228 10:54:40.989609 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.004471 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.006405 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.006889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.010759 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerStarted","Data":"d19859f8e5fd6ed66512dfff8d2897b1d9e4f19dd48408278508fc8015a42b39"} Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.010806 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerStarted","Data":"0e7f82e49b519140b09300846b7a4de854aaea39b1cf5b67bf606b35b6d88c2b"} Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.013572 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.023173 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.038476 4972 generic.go:334] "Generic (PLEG): container finished" podID="4d80893f-182b-4755-afea-fd6635c700fd" containerID="e397650d7d4ff9605e399e9aa184c180ebf09875d4634e14ffedd73ac5d41e04" exitCode=0 Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.040959 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wwvfq" event={"ID":"4d80893f-182b-4755-afea-fd6635c700fd","Type":"ContainerDied","Data":"e397650d7d4ff9605e399e9aa184c180ebf09875d4634e14ffedd73ac5d41e04"} Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.041048 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.041066 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.041482 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.041751 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.062935 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68w4b\" (UniqueName: \"kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b\") pod \"neutron-74884987d7-4szb2\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:41 crc kubenswrapper[4972]: E0228 10:54:41.067929 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d80893f_182b_4755_afea_fd6635c700fd.slice/crio-conmon-e397650d7d4ff9605e399e9aa184c180ebf09875d4634e14ffedd73ac5d41e04.scope\": RecentStats: unable to find data in memory cache]" Feb 28 10:54:41 crc kubenswrapper[4972]: I0228 10:54:41.115557 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:43 crc kubenswrapper[4972]: I0228 10:54:43.064113 4972 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 28 10:54:43 crc kubenswrapper[4972]: I0228 10:54:43.064923 4972 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 28 10:54:43 crc kubenswrapper[4972]: I0228 10:54:43.521141 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:43 crc kubenswrapper[4972]: I0228 10:54:43.600511 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 28 10:54:43 crc kubenswrapper[4972]: I0228 10:54:43.715950 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 28 10:54:44 crc kubenswrapper[4972]: I0228 10:54:44.075079 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.487368 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.635483 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.635965 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.636065 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.636174 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.636205 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.636292 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lvzp\" (UniqueName: \"kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp\") pod \"4d80893f-182b-4755-afea-fd6635c700fd\" (UID: \"4d80893f-182b-4755-afea-fd6635c700fd\") " Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.643147 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.644416 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp" (OuterVolumeSpecName: "kube-api-access-4lvzp") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "kube-api-access-4lvzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.647594 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts" (OuterVolumeSpecName: "scripts") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.649893 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.678624 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data" (OuterVolumeSpecName: "config-data") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.678839 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d80893f-182b-4755-afea-fd6635c700fd" (UID: "4d80893f-182b-4755-afea-fd6635c700fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.737992 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.738024 4972 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.738035 4972 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.738043 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.738052 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d80893f-182b-4755-afea-fd6635c700fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.738064 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lvzp\" (UniqueName: \"kubernetes.io/projected/4d80893f-182b-4755-afea-fd6635c700fd-kube-api-access-4lvzp\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:46 crc kubenswrapper[4972]: W0228 10:54:46.869440 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a53ec32_89b0_4473_a918_cb0d9b104caf.slice/crio-fcbc424c61716e0737789b8a7136e93c2c59bcbfd1a929a95665e9c110b95656 WatchSource:0}: Error finding container fcbc424c61716e0737789b8a7136e93c2c59bcbfd1a929a95665e9c110b95656: Status 404 returned error can't find the container with id fcbc424c61716e0737789b8a7136e93c2c59bcbfd1a929a95665e9c110b95656 Feb 28 10:54:46 crc kubenswrapper[4972]: I0228 10:54:46.870247 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.109318 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerStarted","Data":"fcbc424c61716e0737789b8a7136e93c2c59bcbfd1a929a95665e9c110b95656"} Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.111863 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wwvfq" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.111859 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wwvfq" event={"ID":"4d80893f-182b-4755-afea-fd6635c700fd","Type":"ContainerDied","Data":"ce9560f73438f96621ddeeb6b2ba93ec4e1363bae4c22aca21f064785d4ac0b9"} Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.112025 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce9560f73438f96621ddeeb6b2ba93ec4e1363bae4c22aca21f064785d4ac0b9" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.114035 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerStarted","Data":"b37bfd33de2192e1c15aec2849e6a120e2737b171938309f23f7e3fbad3ea6bc"} Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.114427 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.114852 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.117108 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerStarted","Data":"ea41da9b098e1a55409cb7703684064d876fe2feebbbb5e9891ed25d95639dc8"} Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.143620 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5cf557df54-q4jtg" podStartSLOduration=9.143596967 podStartE2EDuration="9.143596967s" podCreationTimestamp="2026-02-28 10:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:47.141252991 +0000 UTC m=+1144.053260729" watchObservedRunningTime="2026-02-28 10:54:47.143596967 +0000 UTC m=+1144.055604705" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.275658 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.363684 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.363987 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="dnsmasq-dns" containerID="cri-o://4c1f954e8b237c827ce1f2f63600d7eae87ec64c0c2f97136851e301b8e05e55" gracePeriod=10 Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.622013 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-66b556fdbd-nljfl"] Feb 28 10:54:47 crc kubenswrapper[4972]: E0228 10:54:47.623068 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d80893f-182b-4755-afea-fd6635c700fd" containerName="keystone-bootstrap" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.623089 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d80893f-182b-4755-afea-fd6635c700fd" containerName="keystone-bootstrap" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.623421 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d80893f-182b-4755-afea-fd6635c700fd" containerName="keystone-bootstrap" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.626791 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.641776 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9794v" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.641845 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.642027 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.642154 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.642202 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.642357 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.644187 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66b556fdbd-nljfl"] Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759069 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-combined-ca-bundle\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759147 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-public-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759203 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-scripts\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759240 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-credential-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759300 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnb24\" (UniqueName: \"kubernetes.io/projected/cdfe3155-40f6-48df-accc-e7667625cd12-kube-api-access-mnb24\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759335 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-internal-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759364 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-config-data\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.759439 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-fernet-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861573 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-combined-ca-bundle\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861618 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-public-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861660 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-scripts\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861687 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-credential-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861726 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnb24\" (UniqueName: \"kubernetes.io/projected/cdfe3155-40f6-48df-accc-e7667625cd12-kube-api-access-mnb24\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-internal-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861769 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-config-data\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.861831 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-fernet-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.907377 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-internal-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.907778 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnb24\" (UniqueName: \"kubernetes.io/projected/cdfe3155-40f6-48df-accc-e7667625cd12-kube-api-access-mnb24\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.912442 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-combined-ca-bundle\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.915014 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-credential-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.916874 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-public-tls-certs\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.918505 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-config-data\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.918935 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-scripts\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.919115 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cdfe3155-40f6-48df-accc-e7667625cd12-fernet-keys\") pod \"keystone-66b556fdbd-nljfl\" (UID: \"cdfe3155-40f6-48df-accc-e7667625cd12\") " pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:47 crc kubenswrapper[4972]: I0228 10:54:47.947347 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.137889 4972 generic.go:334] "Generic (PLEG): container finished" podID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerID="4c1f954e8b237c827ce1f2f63600d7eae87ec64c0c2f97136851e301b8e05e55" exitCode=0 Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.137982 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" event={"ID":"a622dc9d-f76e-4927-8dde-b0d9dd83220e","Type":"ContainerDied","Data":"4c1f954e8b237c827ce1f2f63600d7eae87ec64c0c2f97136851e301b8e05e55"} Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.143922 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerStarted","Data":"586a361c55d8399ec444105b7af07b9d928fe5a427f041f28aafe2a9736e3763"} Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.383208 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.557562 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6cf8f456-722rs" podUID="a24fd1fd-763d-4106-87b0-805e2f1fc6c5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 28 10:54:48 crc kubenswrapper[4972]: I0228 10:54:48.563241 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-66b556fdbd-nljfl"] Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:48.999897 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.098569 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2kkm\" (UniqueName: \"kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.098632 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.098653 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.098737 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.098934 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.099006 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb\") pod \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\" (UID: \"a622dc9d-f76e-4927-8dde-b0d9dd83220e\") " Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.139875 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm" (OuterVolumeSpecName: "kube-api-access-f2kkm") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "kube-api-access-f2kkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.188369 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerStarted","Data":"191f03e0b519f43c6debe7d7e7ead27033db34c1e41d42193d1dec9559946da1"} Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.190481 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.197810 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66b556fdbd-nljfl" event={"ID":"cdfe3155-40f6-48df-accc-e7667625cd12","Type":"ContainerStarted","Data":"0d47b7b4e1d54224f9ec0c9ecbc2a45cdb0a12e8f0a5d40a6e13ba9708605970"} Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.197840 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-66b556fdbd-nljfl" event={"ID":"cdfe3155-40f6-48df-accc-e7667625cd12","Type":"ContainerStarted","Data":"8700c70be81c7958ef9d84d45ac211f581036dbf167581ca176d08afdfeefe54"} Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.198426 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.209682 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2kkm\" (UniqueName: \"kubernetes.io/projected/a622dc9d-f76e-4927-8dde-b0d9dd83220e-kube-api-access-f2kkm\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.213170 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" event={"ID":"a622dc9d-f76e-4927-8dde-b0d9dd83220e","Type":"ContainerDied","Data":"83b3a3ab8337334c40a871f3625192868a99b8cb6c4419a8cd595bd705311fd5"} Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.213211 4972 scope.go:117] "RemoveContainer" containerID="4c1f954e8b237c827ce1f2f63600d7eae87ec64c0c2f97136851e301b8e05e55" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.213335 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-clq4v" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.226119 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-74884987d7-4szb2" podStartSLOduration=9.22609843 podStartE2EDuration="9.22609843s" podCreationTimestamp="2026-02-28 10:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:49.21783193 +0000 UTC m=+1146.129839668" watchObservedRunningTime="2026-02-28 10:54:49.22609843 +0000 UTC m=+1146.138106168" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.250034 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.306034 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.315953 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.315987 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.323763 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-66b556fdbd-nljfl" podStartSLOduration=2.32374341 podStartE2EDuration="2.32374341s" podCreationTimestamp="2026-02-28 10:54:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:54:49.294081284 +0000 UTC m=+1146.206089032" watchObservedRunningTime="2026-02-28 10:54:49.32374341 +0000 UTC m=+1146.235751148" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.326826 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.354621 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config" (OuterVolumeSpecName: "config") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.419063 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.419113 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.419997 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a622dc9d-f76e-4927-8dde-b0d9dd83220e" (UID: "a622dc9d-f76e-4927-8dde-b0d9dd83220e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.503638 4972 scope.go:117] "RemoveContainer" containerID="afe1436935f6510da9991562f2efe2c1dd01c9f1be8ff0bddf9da2d9fcba1683" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.538100 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a622dc9d-f76e-4927-8dde-b0d9dd83220e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.586520 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.613659 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-clq4v"] Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.676926 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:54:49 crc kubenswrapper[4972]: I0228 10:54:49.827451 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" path="/var/lib/kubelet/pods/a622dc9d-f76e-4927-8dde-b0d9dd83220e/volumes" Feb 28 10:54:50 crc kubenswrapper[4972]: I0228 10:54:50.252775 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtm2c" event={"ID":"aa1673c6-2e2d-496f-b152-ca9df91df6d7","Type":"ContainerStarted","Data":"db4b28d05668240068c901054c9b092d52c2356484af49604f39386d16efdecf"} Feb 28 10:54:50 crc kubenswrapper[4972]: I0228 10:54:50.284102 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qtm2c" podStartSLOduration=2.92742898 podStartE2EDuration="51.284077712s" podCreationTimestamp="2026-02-28 10:53:59 +0000 UTC" firstStartedPulling="2026-02-28 10:54:00.625553343 +0000 UTC m=+1097.537561081" lastFinishedPulling="2026-02-28 10:54:48.982202075 +0000 UTC m=+1145.894209813" observedRunningTime="2026-02-28 10:54:50.274708772 +0000 UTC m=+1147.186716510" watchObservedRunningTime="2026-02-28 10:54:50.284077712 +0000 UTC m=+1147.196085450" Feb 28 10:54:51 crc kubenswrapper[4972]: I0228 10:54:51.288523 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-llqn9" event={"ID":"edd35ce5-02ce-4909-89c5-6dfce27e50a1","Type":"ContainerStarted","Data":"cfda70b6d18e7e219069a3a418223bc19685b3df853a742358b5ed6c6b111ea0"} Feb 28 10:54:51 crc kubenswrapper[4972]: I0228 10:54:51.315773 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-llqn9" podStartSLOduration=4.653144826 podStartE2EDuration="53.315738322s" podCreationTimestamp="2026-02-28 10:53:58 +0000 UTC" firstStartedPulling="2026-02-28 10:54:00.327130429 +0000 UTC m=+1097.239138167" lastFinishedPulling="2026-02-28 10:54:48.989723925 +0000 UTC m=+1145.901731663" observedRunningTime="2026-02-28 10:54:51.308029028 +0000 UTC m=+1148.220036786" watchObservedRunningTime="2026-02-28 10:54:51.315738322 +0000 UTC m=+1148.227746060" Feb 28 10:54:53 crc kubenswrapper[4972]: I0228 10:54:53.318866 4972 generic.go:334] "Generic (PLEG): container finished" podID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" containerID="db4b28d05668240068c901054c9b092d52c2356484af49604f39386d16efdecf" exitCode=0 Feb 28 10:54:53 crc kubenswrapper[4972]: I0228 10:54:53.318943 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtm2c" event={"ID":"aa1673c6-2e2d-496f-b152-ca9df91df6d7","Type":"ContainerDied","Data":"db4b28d05668240068c901054c9b092d52c2356484af49604f39386d16efdecf"} Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.350700 4972 generic.go:334] "Generic (PLEG): container finished" podID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" containerID="cfda70b6d18e7e219069a3a418223bc19685b3df853a742358b5ed6c6b111ea0" exitCode=0 Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.351023 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-llqn9" event={"ID":"edd35ce5-02ce-4909-89c5-6dfce27e50a1","Type":"ContainerDied","Data":"cfda70b6d18e7e219069a3a418223bc19685b3df853a742358b5ed6c6b111ea0"} Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.651150 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.713445 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data\") pod \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.713587 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle\") pod \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.713669 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrtws\" (UniqueName: \"kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws\") pod \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\" (UID: \"aa1673c6-2e2d-496f-b152-ca9df91df6d7\") " Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.721115 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws" (OuterVolumeSpecName: "kube-api-access-zrtws") pod "aa1673c6-2e2d-496f-b152-ca9df91df6d7" (UID: "aa1673c6-2e2d-496f-b152-ca9df91df6d7"). InnerVolumeSpecName "kube-api-access-zrtws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.721132 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "aa1673c6-2e2d-496f-b152-ca9df91df6d7" (UID: "aa1673c6-2e2d-496f-b152-ca9df91df6d7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.745962 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa1673c6-2e2d-496f-b152-ca9df91df6d7" (UID: "aa1673c6-2e2d-496f-b152-ca9df91df6d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.816239 4972 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.816272 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa1673c6-2e2d-496f-b152-ca9df91df6d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:56 crc kubenswrapper[4972]: I0228 10:54:56.816282 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrtws\" (UniqueName: \"kubernetes.io/projected/aa1673c6-2e2d-496f-b152-ca9df91df6d7-kube-api-access-zrtws\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.365282 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qtm2c" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.367612 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qtm2c" event={"ID":"aa1673c6-2e2d-496f-b152-ca9df91df6d7","Type":"ContainerDied","Data":"22ecb3ffeb62e6b9938c8245ef2c211afd9d9665f95b2a6e070c4db7eaac3b23"} Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.367667 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22ecb3ffeb62e6b9938c8245ef2c211afd9d9665f95b2a6e070c4db7eaac3b23" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.850954 4972 scope.go:117] "RemoveContainer" containerID="625df873be954ef27d6136bf8de206a89bf44bfaec7dd2f36cd25ca732df513a" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.987594 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-64d46dc98-5dqwn"] Feb 28 10:54:57 crc kubenswrapper[4972]: E0228 10:54:57.988052 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="dnsmasq-dns" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.988072 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="dnsmasq-dns" Feb 28 10:54:57 crc kubenswrapper[4972]: E0228 10:54:57.988087 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="init" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.988093 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="init" Feb 28 10:54:57 crc kubenswrapper[4972]: E0228 10:54:57.988117 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" containerName="barbican-db-sync" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.988123 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" containerName="barbican-db-sync" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.988333 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" containerName="barbican-db-sync" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.988352 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a622dc9d-f76e-4927-8dde-b0d9dd83220e" containerName="dnsmasq-dns" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.989418 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.993983 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-k9jst" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.994216 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 28 10:54:57 crc kubenswrapper[4972]: I0228 10:54:57.994378 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.011354 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64d46dc98-5dqwn"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.021218 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5d9686d86c-lmjsc"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.024575 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.028104 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.056930 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data-custom\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.057113 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.057223 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsc8x\" (UniqueName: \"kubernetes.io/projected/fed4da99-54b6-4444-8f0f-4489a5286e4d-kube-api-access-jsc8x\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.057386 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-combined-ca-bundle\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.057498 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed4da99-54b6-4444-8f0f-4489a5286e4d-logs\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.075582 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9686d86c-lmjsc"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159421 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj5wv\" (UniqueName: \"kubernetes.io/projected/31309c10-9f68-4aed-ab99-dcef67e12b92-kube-api-access-vj5wv\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159476 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-combined-ca-bundle\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159512 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsc8x\" (UniqueName: \"kubernetes.io/projected/fed4da99-54b6-4444-8f0f-4489a5286e4d-kube-api-access-jsc8x\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159562 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data-custom\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159597 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-combined-ca-bundle\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159621 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159654 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed4da99-54b6-4444-8f0f-4489a5286e4d-logs\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159694 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data-custom\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159725 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31309c10-9f68-4aed-ab99-dcef67e12b92-logs\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.159760 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.164309 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed4da99-54b6-4444-8f0f-4489a5286e4d-logs\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.168187 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-combined-ca-bundle\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.170183 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data-custom\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.175812 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.176040 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed4da99-54b6-4444-8f0f-4489a5286e4d-config-data\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.177516 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.189133 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsc8x\" (UniqueName: \"kubernetes.io/projected/fed4da99-54b6-4444-8f0f-4489a5286e4d-kube-api-access-jsc8x\") pod \"barbican-keystone-listener-64d46dc98-5dqwn\" (UID: \"fed4da99-54b6-4444-8f0f-4489a5286e4d\") " pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.193419 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263669 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263725 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263756 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj5wv\" (UniqueName: \"kubernetes.io/projected/31309c10-9f68-4aed-ab99-dcef67e12b92-kube-api-access-vj5wv\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263778 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgskb\" (UniqueName: \"kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-combined-ca-bundle\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263845 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263871 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data-custom\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263895 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263921 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263942 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.263992 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31309c10-9f68-4aed-ab99-dcef67e12b92-logs\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.264430 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31309c10-9f68-4aed-ab99-dcef67e12b92-logs\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.270715 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.272056 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-llqn9" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.273101 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-config-data-custom\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.279138 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31309c10-9f68-4aed-ab99-dcef67e12b92-combined-ca-bundle\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.291232 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:54:58 crc kubenswrapper[4972]: E0228 10:54:58.291678 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" containerName="cinder-db-sync" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.291696 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" containerName="cinder-db-sync" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.291861 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" containerName="cinder-db-sync" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.292865 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.304713 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.313149 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.326436 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj5wv\" (UniqueName: \"kubernetes.io/projected/31309c10-9f68-4aed-ab99-dcef67e12b92-kube-api-access-vj5wv\") pod \"barbican-worker-5d9686d86c-lmjsc\" (UID: \"31309c10-9f68-4aed-ab99-dcef67e12b92\") " pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.333131 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.366418 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9686d86c-lmjsc" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367245 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367388 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367490 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367579 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367618 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367680 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25ds5\" (UniqueName: \"kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5\") pod \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\" (UID: \"edd35ce5-02ce-4909-89c5-6dfce27e50a1\") " Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367919 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367953 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.367992 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368022 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368052 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368117 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368134 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368154 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368169 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368196 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgskb\" (UniqueName: \"kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.368225 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rcz\" (UniqueName: \"kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.374613 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.376543 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.376724 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.377246 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.382797 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.384624 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.385848 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.386734 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts" (OuterVolumeSpecName: "scripts") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.395212 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5" (OuterVolumeSpecName: "kube-api-access-25ds5") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "kube-api-access-25ds5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.404142 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.414134 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgskb\" (UniqueName: \"kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb\") pod \"dnsmasq-dns-85ff748b95-b89lq\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.417317 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.430010 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-llqn9" event={"ID":"edd35ce5-02ce-4909-89c5-6dfce27e50a1","Type":"ContainerDied","Data":"5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09"} Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.430083 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a3ec62214fdf4212110335361bc4750fe8e84125e23cf3da902ffbf6251ce09" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.430173 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-llqn9" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.456633 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data" (OuterVolumeSpecName: "config-data") pod "edd35ce5-02ce-4909-89c5-6dfce27e50a1" (UID: "edd35ce5-02ce-4909-89c5-6dfce27e50a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472677 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472725 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472774 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rcz\" (UniqueName: \"kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472859 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472897 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472978 4972 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.472989 4972 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd35ce5-02ce-4909-89c5-6dfce27e50a1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.473000 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25ds5\" (UniqueName: \"kubernetes.io/projected/edd35ce5-02ce-4909-89c5-6dfce27e50a1-kube-api-access-25ds5\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.473010 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.473018 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.473028 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd35ce5-02ce-4909-89c5-6dfce27e50a1-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.476857 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.479806 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.480349 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.481641 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.506136 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rcz\" (UniqueName: \"kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz\") pod \"barbican-api-7f59c67b56-sft76\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.556209 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6cf8f456-722rs" podUID="a24fd1fd-763d-4106-87b0-805e2f1fc6c5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.635076 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.655205 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.795518 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.804073 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.814721 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rm7g4" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.847017 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.850638 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.851009 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.871566 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.906741 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.906835 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.906900 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.907890 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.907945 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80" gracePeriod=600 Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.923010 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.923163 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q89c\" (UniqueName: \"kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.923311 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.923543 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.926198 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.926288 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:58 crc kubenswrapper[4972]: I0228 10:54:58.949535 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.008615 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.010630 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.020640 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033642 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033736 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033763 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033814 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q89c\" (UniqueName: \"kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.033956 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.043569 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.058038 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.059014 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.071246 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.078287 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q89c\" (UniqueName: \"kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.085084 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts\") pod \"cinder-scheduler-0\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.135894 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.135987 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t858g\" (UniqueName: \"kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.136035 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.136076 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.136173 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.136214 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.182225 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.189071 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.196881 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.211493 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.229438 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238434 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238512 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238581 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238607 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238651 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.238691 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t858g\" (UniqueName: \"kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.246111 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.246836 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.247912 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.248575 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.249399 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.268288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t858g\" (UniqueName: \"kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g\") pod \"dnsmasq-dns-5c9776ccc5-n4z6j\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.340986 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341045 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341078 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341136 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4whgt\" (UniqueName: \"kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341213 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341234 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.341298 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.383239 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-64d46dc98-5dqwn"] Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.442885 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443410 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443434 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443504 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4whgt\" (UniqueName: \"kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443565 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443586 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.443623 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.446514 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.446937 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: W0228 10:54:59.455617 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfed4da99_54b6_4444_8f0f_4489a5286e4d.slice/crio-c849ebf413f6c7b8015be537b575fe7676e8c12351be32f53ad059c7c9a8fc68 WatchSource:0}: Error finding container c849ebf413f6c7b8015be537b575fe7676e8c12351be32f53ad059c7c9a8fc68: Status 404 returned error can't find the container with id c849ebf413f6c7b8015be537b575fe7676e8c12351be32f53ad059c7c9a8fc68 Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.457208 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.457602 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.459163 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.465034 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.479252 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80" exitCode=0 Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.479325 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80"} Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.479378 4972 scope.go:117] "RemoveContainer" containerID="aaa26c39239d3630d84e46cefa875b954a7725731ff56c79ba02a6d7067261e3" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.495064 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4whgt\" (UniqueName: \"kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt\") pod \"cinder-api-0\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.512248 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.576667 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:54:59 crc kubenswrapper[4972]: E0228 10:54:59.602919 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" Feb 28 10:54:59 crc kubenswrapper[4972]: I0228 10:54:59.627224 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9686d86c-lmjsc"] Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.210121 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.243356 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.258367 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.343862 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.348187 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:00 crc kubenswrapper[4972]: W0228 10:55:00.370867 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1f9fb87_9896_48e6_9491_24b4a96ec91a.slice/crio-98550b6eac9a707fc12adae4fdd1b52357d3b67644d4d79e0e4434d030de1500 WatchSource:0}: Error finding container 98550b6eac9a707fc12adae4fdd1b52357d3b67644d4d79e0e4434d030de1500: Status 404 returned error can't find the container with id 98550b6eac9a707fc12adae4fdd1b52357d3b67644d4d79e0e4434d030de1500 Feb 28 10:55:00 crc kubenswrapper[4972]: W0228 10:55:00.377816 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9499375a_dad8_43ee_af2f_02963356cf50.slice/crio-24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487 WatchSource:0}: Error finding container 24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487: Status 404 returned error can't find the container with id 24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487 Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.506052 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerStarted","Data":"98550b6eac9a707fc12adae4fdd1b52357d3b67644d4d79e0e4434d030de1500"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.520197 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" event={"ID":"fed4da99-54b6-4444-8f0f-4489a5286e4d","Type":"ContainerStarted","Data":"c849ebf413f6c7b8015be537b575fe7676e8c12351be32f53ad059c7c9a8fc68"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.523411 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9686d86c-lmjsc" event={"ID":"31309c10-9f68-4aed-ab99-dcef67e12b92","Type":"ContainerStarted","Data":"16e3b534fd30f4b53643698eaa1f9edfdef5790897d2dc633bba0ef058a11da0"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.533760 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerStarted","Data":"493d8e0aac96982d094ce0ff886ad15997d2500e77a72ec1a782835ca8900db5"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.533901 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="ceilometer-notification-agent" containerID="cri-o://9b5761dfc835d28b55d4bd522cef3703451247ab368d78fd87ea8706db0cec97" gracePeriod=30 Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.534025 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.534331 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="sg-core" containerID="cri-o://ea41da9b098e1a55409cb7703684064d876fe2feebbbb5e9891ed25d95639dc8" gracePeriod=30 Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.534406 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="proxy-httpd" containerID="cri-o://493d8e0aac96982d094ce0ff886ad15997d2500e77a72ec1a782835ca8900db5" gracePeriod=30 Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.540899 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.544835 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerStarted","Data":"7a870468f66ed0384580066c7ba9097e41693def265b34ac871f56d9aba3620e"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.546663 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" event={"ID":"0f9cfa71-3153-407d-82a1-d92103b6d681","Type":"ContainerStarted","Data":"d7a88684431ee80e868de2a90dcb9494af62d5264938b26deb4982f3e2eb8e49"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.548634 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" event={"ID":"9499375a-dad8-43ee-af2f-02963356cf50","Type":"ContainerStarted","Data":"24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487"} Feb 28 10:55:00 crc kubenswrapper[4972]: I0228 10:55:00.551361 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerStarted","Data":"fa7e9ac5f87b3f1424d5a787ce425d686b90ef4ba870f5ce2b8028f738746685"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.663242 4972 generic.go:334] "Generic (PLEG): container finished" podID="9499375a-dad8-43ee-af2f-02963356cf50" containerID="123f1fa701904d68ab41ba1c014c5b182c037442687f66e541af8aeb25cb1959" exitCode=0 Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.663937 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" event={"ID":"9499375a-dad8-43ee-af2f-02963356cf50","Type":"ContainerDied","Data":"123f1fa701904d68ab41ba1c014c5b182c037442687f66e541af8aeb25cb1959"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.689010 4972 generic.go:334] "Generic (PLEG): container finished" podID="84e043f8-8906-4baf-a610-53446e9c0f85" containerID="493d8e0aac96982d094ce0ff886ad15997d2500e77a72ec1a782835ca8900db5" exitCode=0 Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.689038 4972 generic.go:334] "Generic (PLEG): container finished" podID="84e043f8-8906-4baf-a610-53446e9c0f85" containerID="ea41da9b098e1a55409cb7703684064d876fe2feebbbb5e9891ed25d95639dc8" exitCode=2 Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.689080 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerDied","Data":"493d8e0aac96982d094ce0ff886ad15997d2500e77a72ec1a782835ca8900db5"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.689109 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerDied","Data":"ea41da9b098e1a55409cb7703684064d876fe2feebbbb5e9891ed25d95639dc8"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.769693 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerStarted","Data":"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.770120 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerStarted","Data":"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.770285 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.770717 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.807938 4972 generic.go:334] "Generic (PLEG): container finished" podID="0f9cfa71-3153-407d-82a1-d92103b6d681" containerID="235469bee88e34ed4da8757350b27b8643aeaa0bedfda01fb4d5e06d425246ee" exitCode=0 Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.853768 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" event={"ID":"0f9cfa71-3153-407d-82a1-d92103b6d681","Type":"ContainerDied","Data":"235469bee88e34ed4da8757350b27b8643aeaa0bedfda01fb4d5e06d425246ee"} Feb 28 10:55:01 crc kubenswrapper[4972]: I0228 10:55:01.918683 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f59c67b56-sft76" podStartSLOduration=3.918659162 podStartE2EDuration="3.918659162s" podCreationTimestamp="2026-02-28 10:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:01.852415518 +0000 UTC m=+1158.764423256" watchObservedRunningTime="2026-02-28 10:55:01.918659162 +0000 UTC m=+1158.830666920" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.530404 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.829115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerStarted","Data":"1e3eab8fad2e2ca159b4522f575dc0edaa0f66decc9afccb3fa57e3c5d2134a4"} Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.835299 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" event={"ID":"0f9cfa71-3153-407d-82a1-d92103b6d681","Type":"ContainerDied","Data":"d7a88684431ee80e868de2a90dcb9494af62d5264938b26deb4982f3e2eb8e49"} Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.835347 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7a88684431ee80e868de2a90dcb9494af62d5264938b26deb4982f3e2eb8e49" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.835801 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.842388 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerStarted","Data":"3d7ff4abdbc1386075be8c3d6987ed3d6a21b6bee82abc73cef203061779a37f"} Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.937927 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.938023 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgskb\" (UniqueName: \"kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.938067 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.938085 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.938172 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.938230 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0\") pod \"0f9cfa71-3153-407d-82a1-d92103b6d681\" (UID: \"0f9cfa71-3153-407d-82a1-d92103b6d681\") " Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.948287 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb" (OuterVolumeSpecName: "kube-api-access-pgskb") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "kube-api-access-pgskb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.965619 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.982059 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config" (OuterVolumeSpecName: "config") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:02 crc kubenswrapper[4972]: I0228 10:55:02.988318 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.005601 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.013127 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f9cfa71-3153-407d-82a1-d92103b6d681" (UID: "0f9cfa71-3153-407d-82a1-d92103b6d681"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.046635 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.046905 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgskb\" (UniqueName: \"kubernetes.io/projected/0f9cfa71-3153-407d-82a1-d92103b6d681-kube-api-access-pgskb\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.047635 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.047655 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.047668 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.047679 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f9cfa71-3153-407d-82a1-d92103b6d681-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.873219 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-b89lq" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.874487 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" event={"ID":"9499375a-dad8-43ee-af2f-02963356cf50","Type":"ContainerStarted","Data":"bf541cc76cf02c659b3a50beeae0508bd468413e17e045507719169ba8bc0733"} Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.877119 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:55:03 crc kubenswrapper[4972]: I0228 10:55:03.927302 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" podStartSLOduration=5.927278047 podStartE2EDuration="5.927278047s" podCreationTimestamp="2026-02-28 10:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:03.922040131 +0000 UTC m=+1160.834047869" watchObservedRunningTime="2026-02-28 10:55:03.927278047 +0000 UTC m=+1160.839285785" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:03.995661 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.002595 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-b89lq"] Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.885594 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerStarted","Data":"cc36b1309774da247b91f92e933c771ce7b01ef579a7b1423d5110aab679d5c7"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.886124 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.885842 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api" containerID="cri-o://cc36b1309774da247b91f92e933c771ce7b01ef579a7b1423d5110aab679d5c7" gracePeriod=30 Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.885721 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api-log" containerID="cri-o://1e3eab8fad2e2ca159b4522f575dc0edaa0f66decc9afccb3fa57e3c5d2134a4" gracePeriod=30 Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.889510 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" event={"ID":"fed4da99-54b6-4444-8f0f-4489a5286e4d","Type":"ContainerStarted","Data":"34742fdc1dc4b242cb1824b37c31fa8a73a6e5c27fc8b6bfd43661311c5e66bc"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.889593 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" event={"ID":"fed4da99-54b6-4444-8f0f-4489a5286e4d","Type":"ContainerStarted","Data":"1755668afb5e9b26d1edf18046fcbd368968116a345eda0d9879ecd5bc1cdd5e"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.894123 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9686d86c-lmjsc" event={"ID":"31309c10-9f68-4aed-ab99-dcef67e12b92","Type":"ContainerStarted","Data":"d3b3eb45a3d7dd9236b7052affbabd6af72c7b3d5ca2bb91e0d8f11d64f5365d"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.894190 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9686d86c-lmjsc" event={"ID":"31309c10-9f68-4aed-ab99-dcef67e12b92","Type":"ContainerStarted","Data":"61616ac58edcaf6c076df12f88761aba527c89137fda4ad257a74b7cbde628e1"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.897667 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerStarted","Data":"75ae2c7c3c6937c7635798c50df53d6845872507f4ba65acae18046ebb3f727f"} Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.914032 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.914008196 podStartE2EDuration="5.914008196s" podCreationTimestamp="2026-02-28 10:54:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:04.911992529 +0000 UTC m=+1161.824000257" watchObservedRunningTime="2026-02-28 10:55:04.914008196 +0000 UTC m=+1161.826015934" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.944394 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9f94fccfd-62888"] Feb 28 10:55:04 crc kubenswrapper[4972]: E0228 10:55:04.945185 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9cfa71-3153-407d-82a1-d92103b6d681" containerName="init" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.945205 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9cfa71-3153-407d-82a1-d92103b6d681" containerName="init" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.945409 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f9cfa71-3153-407d-82a1-d92103b6d681" containerName="init" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.946448 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.949303 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.949609 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.959320 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.94856465 podStartE2EDuration="6.959298157s" podCreationTimestamp="2026-02-28 10:54:58 +0000 UTC" firstStartedPulling="2026-02-28 10:55:00.242643833 +0000 UTC m=+1157.154651571" lastFinishedPulling="2026-02-28 10:55:01.25337733 +0000 UTC m=+1158.165385078" observedRunningTime="2026-02-28 10:55:04.945108302 +0000 UTC m=+1161.857116050" watchObservedRunningTime="2026-02-28 10:55:04.959298157 +0000 UTC m=+1161.871305895" Feb 28 10:55:04 crc kubenswrapper[4972]: I0228 10:55:04.964756 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f94fccfd-62888"] Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.000060 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-64d46dc98-5dqwn" podStartSLOduration=3.975899279 podStartE2EDuration="8.000035382s" podCreationTimestamp="2026-02-28 10:54:57 +0000 UTC" firstStartedPulling="2026-02-28 10:54:59.495318354 +0000 UTC m=+1156.407326092" lastFinishedPulling="2026-02-28 10:55:03.519454457 +0000 UTC m=+1160.431462195" observedRunningTime="2026-02-28 10:55:04.997298605 +0000 UTC m=+1161.909306343" watchObservedRunningTime="2026-02-28 10:55:05.000035382 +0000 UTC m=+1161.912043130" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009330 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-public-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009414 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wzzx\" (UniqueName: \"kubernetes.io/projected/e50121ef-8724-4a1d-8ee2-badc2c40df58-kube-api-access-8wzzx\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009482 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009626 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-internal-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009780 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data-custom\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009831 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-combined-ca-bundle\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.009967 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50121ef-8724-4a1d-8ee2-badc2c40df58-logs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.100021 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5d9686d86c-lmjsc" podStartSLOduration=4.222363706 podStartE2EDuration="8.099561305s" podCreationTimestamp="2026-02-28 10:54:57 +0000 UTC" firstStartedPulling="2026-02-28 10:54:59.644352226 +0000 UTC m=+1156.556359964" lastFinishedPulling="2026-02-28 10:55:03.521549825 +0000 UTC m=+1160.433557563" observedRunningTime="2026-02-28 10:55:05.034824351 +0000 UTC m=+1161.946832089" watchObservedRunningTime="2026-02-28 10:55:05.099561305 +0000 UTC m=+1162.011569043" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.111906 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50121ef-8724-4a1d-8ee2-badc2c40df58-logs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112019 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-public-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112047 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wzzx\" (UniqueName: \"kubernetes.io/projected/e50121ef-8724-4a1d-8ee2-badc2c40df58-kube-api-access-8wzzx\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112076 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112138 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-internal-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112209 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data-custom\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.112237 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-combined-ca-bundle\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.114147 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e50121ef-8724-4a1d-8ee2-badc2c40df58-logs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.122237 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-public-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.122342 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-internal-tls-certs\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.127626 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data-custom\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.128333 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-combined-ca-bundle\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.140889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wzzx\" (UniqueName: \"kubernetes.io/projected/e50121ef-8724-4a1d-8ee2-badc2c40df58-kube-api-access-8wzzx\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.145998 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50121ef-8724-4a1d-8ee2-badc2c40df58-config-data\") pod \"barbican-api-9f94fccfd-62888\" (UID: \"e50121ef-8724-4a1d-8ee2-badc2c40df58\") " pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.378642 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.808698 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f9cfa71-3153-407d-82a1-d92103b6d681" path="/var/lib/kubelet/pods/0f9cfa71-3153-407d-82a1-d92103b6d681/volumes" Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.927326 4972 generic.go:334] "Generic (PLEG): container finished" podID="84e043f8-8906-4baf-a610-53446e9c0f85" containerID="9b5761dfc835d28b55d4bd522cef3703451247ab368d78fd87ea8706db0cec97" exitCode=0 Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.927390 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerDied","Data":"9b5761dfc835d28b55d4bd522cef3703451247ab368d78fd87ea8706db0cec97"} Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.928723 4972 generic.go:334] "Generic (PLEG): container finished" podID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerID="1e3eab8fad2e2ca159b4522f575dc0edaa0f66decc9afccb3fa57e3c5d2134a4" exitCode=143 Feb 28 10:55:05 crc kubenswrapper[4972]: I0228 10:55:05.932670 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerDied","Data":"1e3eab8fad2e2ca159b4522f575dc0edaa0f66decc9afccb3fa57e3c5d2134a4"} Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.033475 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f94fccfd-62888"] Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.190376 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236155 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236253 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5tlh\" (UniqueName: \"kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236292 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236413 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236445 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236524 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.236619 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.238757 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.238807 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.245772 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh" (OuterVolumeSpecName: "kube-api-access-t5tlh") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "kube-api-access-t5tlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.251490 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts" (OuterVolumeSpecName: "scripts") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.282113 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: E0228 10:55:06.334904 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle podName:84e043f8-8906-4baf-a610-53446e9c0f85 nodeName:}" failed. No retries permitted until 2026-02-28 10:55:06.834870657 +0000 UTC m=+1163.746878405 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85") : error deleting /var/lib/kubelet/pods/84e043f8-8906-4baf-a610-53446e9c0f85/volume-subpaths: remove /var/lib/kubelet/pods/84e043f8-8906-4baf-a610-53446e9c0f85/volume-subpaths: no such file or directory Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339127 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339154 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339164 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339174 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5tlh\" (UniqueName: \"kubernetes.io/projected/84e043f8-8906-4baf-a610-53446e9c0f85-kube-api-access-t5tlh\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339186 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84e043f8-8906-4baf-a610-53446e9c0f85-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.339324 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data" (OuterVolumeSpecName: "config-data") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.441604 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.849857 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") pod \"84e043f8-8906-4baf-a610-53446e9c0f85\" (UID: \"84e043f8-8906-4baf-a610-53446e9c0f85\") " Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.864220 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84e043f8-8906-4baf-a610-53446e9c0f85" (UID: "84e043f8-8906-4baf-a610-53446e9c0f85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.944187 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.944152 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84e043f8-8906-4baf-a610-53446e9c0f85","Type":"ContainerDied","Data":"944e6d267bdb266a5a395de936b90fc9553f9b8d9ec3e1898f8995ae3d9c5abd"} Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.944321 4972 scope.go:117] "RemoveContainer" containerID="493d8e0aac96982d094ce0ff886ad15997d2500e77a72ec1a782835ca8900db5" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.949803 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f94fccfd-62888" event={"ID":"e50121ef-8724-4a1d-8ee2-badc2c40df58","Type":"ContainerStarted","Data":"c62bad0b5ae4c56f05859ebe1223bdbefaa417cb9ed853642d92787c9535461a"} Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.950033 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f94fccfd-62888" event={"ID":"e50121ef-8724-4a1d-8ee2-badc2c40df58","Type":"ContainerStarted","Data":"1fea29f1afdb0af0d656132995ffe017e62c4907387a01bc31cb2e2d2cabb1c6"} Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.950048 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f94fccfd-62888" event={"ID":"e50121ef-8724-4a1d-8ee2-badc2c40df58","Type":"ContainerStarted","Data":"fc6221da35b0bd670d67972881f9f546b46060a770dce63b6ec0186457fe46d0"} Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.950143 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.950341 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.953715 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e043f8-8906-4baf-a610-53446e9c0f85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.989065 4972 scope.go:117] "RemoveContainer" containerID="ea41da9b098e1a55409cb7703684064d876fe2feebbbb5e9891ed25d95639dc8" Feb 28 10:55:06 crc kubenswrapper[4972]: I0228 10:55:06.989885 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9f94fccfd-62888" podStartSLOduration=2.989862804 podStartE2EDuration="2.989862804s" podCreationTimestamp="2026-02-28 10:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:06.978043134 +0000 UTC m=+1163.890050872" watchObservedRunningTime="2026-02-28 10:55:06.989862804 +0000 UTC m=+1163.901870542" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.024808 4972 scope.go:117] "RemoveContainer" containerID="9b5761dfc835d28b55d4bd522cef3703451247ab368d78fd87ea8706db0cec97" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.067540 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.091530 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107159 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:07 crc kubenswrapper[4972]: E0228 10:55:07.107606 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="sg-core" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107622 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="sg-core" Feb 28 10:55:07 crc kubenswrapper[4972]: E0228 10:55:07.107638 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="proxy-httpd" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107644 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="proxy-httpd" Feb 28 10:55:07 crc kubenswrapper[4972]: E0228 10:55:07.107669 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="ceilometer-notification-agent" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107688 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="ceilometer-notification-agent" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107882 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="ceilometer-notification-agent" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107906 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="proxy-httpd" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.107918 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" containerName="sg-core" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.110074 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.117023 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.117175 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.124219 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165726 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165780 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165827 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165870 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165912 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5p77\" (UniqueName: \"kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165952 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.165980 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.267933 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268016 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268062 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5p77\" (UniqueName: \"kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268106 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268137 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268492 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.268706 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.269064 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.269099 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.273078 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.273202 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.273515 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.275881 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.287616 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5p77\" (UniqueName: \"kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77\") pod \"ceilometer-0\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.367488 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.436185 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.757993 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.759069 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74884987d7-4szb2" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-api" containerID="cri-o://586a361c55d8399ec444105b7af07b9d928fe5a427f041f28aafe2a9736e3763" gracePeriod=30 Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.759746 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-74884987d7-4szb2" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-httpd" containerID="cri-o://191f03e0b519f43c6debe7d7e7ead27033db34c1e41d42193d1dec9559946da1" gracePeriod=30 Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.769934 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69dd77478c-x64n9"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.778567 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dd77478c-x64n9"] Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.772011 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.852244 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e043f8-8906-4baf-a610-53446e9c0f85" path="/var/lib/kubelet/pods/84e043f8-8906-4baf-a610-53446e9c0f85/volumes" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.853185 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.888632 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-ovndb-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.888736 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-internal-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.888837 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-combined-ca-bundle\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.888870 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.888973 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-public-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.889018 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljvnx\" (UniqueName: \"kubernetes.io/projected/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-kube-api-access-ljvnx\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.889050 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-httpd-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.996129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-ovndb-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.996658 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-internal-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.996935 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-combined-ca-bundle\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.996969 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.997038 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-public-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.997073 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljvnx\" (UniqueName: \"kubernetes.io/projected/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-kube-api-access-ljvnx\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:07 crc kubenswrapper[4972]: I0228 10:55:07.997109 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-httpd-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.006138 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-httpd-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.007195 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-internal-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.008154 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-ovndb-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.008225 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-public-tls-certs\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.011347 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-config\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.016916 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.025314 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-combined-ca-bundle\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.032257 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljvnx\" (UniqueName: \"kubernetes.io/projected/b7d2ee5d-2b90-4c97-a931-45992f1f50d1-kube-api-access-ljvnx\") pod \"neutron-69dd77478c-x64n9\" (UID: \"b7d2ee5d-2b90-4c97-a931-45992f1f50d1\") " pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.144366 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:08 crc kubenswrapper[4972]: I0228 10:55:08.831259 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69dd77478c-x64n9"] Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.003406 4972 generic.go:334] "Generic (PLEG): container finished" podID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerID="191f03e0b519f43c6debe7d7e7ead27033db34c1e41d42193d1dec9559946da1" exitCode=0 Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.003499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerDied","Data":"191f03e0b519f43c6debe7d7e7ead27033db34c1e41d42193d1dec9559946da1"} Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.006264 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerStarted","Data":"51f1fc087d80d26d52cf588d5399be26f43b07d01508a23524514db3f82a113b"} Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.016399 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd77478c-x64n9" event={"ID":"b7d2ee5d-2b90-4c97-a931-45992f1f50d1","Type":"ContainerStarted","Data":"6987da98b25d27d474d80ebc0fa9a0c2715fb918353de48b952175498a746035"} Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.232298 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.530612 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.579858 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.652703 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:55:09 crc kubenswrapper[4972]: I0228 10:55:09.653651 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="dnsmasq-dns" containerID="cri-o://202f19364f0f0b5b5babf7a2d417cfe899433378faf2c7d685894c7929a8cce0" gracePeriod=10 Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.032941 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerStarted","Data":"7925996f825b8a1aef278517beec9a3e279ad44c1f465e886cd5bcff6dfcc881"} Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.037544 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd77478c-x64n9" event={"ID":"b7d2ee5d-2b90-4c97-a931-45992f1f50d1","Type":"ContainerStarted","Data":"6ada9e5163d2704bf04327660480cc77cee79c1eee12541574d1663e129f16e8"} Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.037582 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69dd77478c-x64n9" event={"ID":"b7d2ee5d-2b90-4c97-a931-45992f1f50d1","Type":"ContainerStarted","Data":"be3ab347133ac07335a18d469aa67ed0184c45c623f62beab10cc6d4dd812ba6"} Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.038847 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.056724 4972 generic.go:334] "Generic (PLEG): container finished" podID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerID="202f19364f0f0b5b5babf7a2d417cfe899433378faf2c7d685894c7929a8cce0" exitCode=0 Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.056825 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" event={"ID":"0bc05540-b9af-4d90-abd7-2b6d4d8ef976","Type":"ContainerDied","Data":"202f19364f0f0b5b5babf7a2d417cfe899433378faf2c7d685894c7929a8cce0"} Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.152284 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69dd77478c-x64n9" podStartSLOduration=3.152262768 podStartE2EDuration="3.152262768s" podCreationTimestamp="2026-02-28 10:55:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:10.103811455 +0000 UTC m=+1167.015819193" watchObservedRunningTime="2026-02-28 10:55:10.152262768 +0000 UTC m=+1167.064270506" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.258961 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.550432 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.717507 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.717604 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.717833 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.717996 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.718057 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.718095 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjls5\" (UniqueName: \"kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5\") pod \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\" (UID: \"0bc05540-b9af-4d90-abd7-2b6d4d8ef976\") " Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.736619 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5" (OuterVolumeSpecName: "kube-api-access-mjls5") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "kube-api-access-mjls5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.810567 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.820677 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjls5\" (UniqueName: \"kubernetes.io/projected/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-kube-api-access-mjls5\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.820719 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.841145 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.864093 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config" (OuterVolumeSpecName: "config") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.865581 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.914355 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0bc05540-b9af-4d90-abd7-2b6d4d8ef976" (UID: "0bc05540-b9af-4d90-abd7-2b6d4d8ef976"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.923876 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.923931 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.923945 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:10 crc kubenswrapper[4972]: I0228 10:55:10.923957 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bc05540-b9af-4d90-abd7-2b6d4d8ef976-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.069889 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerStarted","Data":"64497dc29412d515d95a37308ed7726b923f50c4fd91c390e346dc327a49c600"} Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.081978 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="cinder-scheduler" containerID="cri-o://3d7ff4abdbc1386075be8c3d6987ed3d6a21b6bee82abc73cef203061779a37f" gracePeriod=30 Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.082422 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="probe" containerID="cri-o://75ae2c7c3c6937c7635798c50df53d6845872507f4ba65acae18046ebb3f727f" gracePeriod=30 Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.082573 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" event={"ID":"0bc05540-b9af-4d90-abd7-2b6d4d8ef976","Type":"ContainerDied","Data":"fee32373a4e5c66360bc1856106c63c7b5e14c9ae94bdf4846f53385c32849ae"} Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.082624 4972 scope.go:117] "RemoveContainer" containerID="202f19364f0f0b5b5babf7a2d417cfe899433378faf2c7d685894c7929a8cce0" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.082522 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-b5sjv" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.121244 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-74884987d7-4szb2" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.160:9696/\": dial tcp 10.217.0.160:9696: connect: connection refused" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.201806 4972 scope.go:117] "RemoveContainer" containerID="bf116311e059a34e36dcb51d90e94848584b74ec778f7dc7aa3afcc4af7116f1" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.211644 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.232046 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-b5sjv"] Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.496033 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.832898 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" path="/var/lib/kubelet/pods/0bc05540-b9af-4d90-abd7-2b6d4d8ef976/volumes" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.833762 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5bb67f8988-tt9dr"] Feb 28 10:55:11 crc kubenswrapper[4972]: E0228 10:55:11.834108 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="init" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.834126 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="init" Feb 28 10:55:11 crc kubenswrapper[4972]: E0228 10:55:11.834155 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="dnsmasq-dns" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.834162 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="dnsmasq-dns" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.834372 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc05540-b9af-4d90-abd7-2b6d4d8ef976" containerName="dnsmasq-dns" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.835389 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5bb67f8988-tt9dr"] Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.838034 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960616 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae387b2-3007-4c94-9a6f-255bd137d7c0-logs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960804 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6tx\" (UniqueName: \"kubernetes.io/projected/bae387b2-3007-4c94-9a6f-255bd137d7c0-kube-api-access-rw6tx\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960837 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-internal-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960888 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-combined-ca-bundle\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960910 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-config-data\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960930 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-scripts\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:11 crc kubenswrapper[4972]: I0228 10:55:11.960951 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-public-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.065842 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-combined-ca-bundle\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.065892 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-config-data\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.065919 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-scripts\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.065948 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-public-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.065978 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae387b2-3007-4c94-9a6f-255bd137d7c0-logs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.066216 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6tx\" (UniqueName: \"kubernetes.io/projected/bae387b2-3007-4c94-9a6f-255bd137d7c0-kube-api-access-rw6tx\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.066263 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-internal-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.067508 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae387b2-3007-4c94-9a6f-255bd137d7c0-logs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.073315 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-config-data\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.077161 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-public-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.077716 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-scripts\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.081064 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-internal-tls-certs\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.084999 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae387b2-3007-4c94-9a6f-255bd137d7c0-combined-ca-bundle\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.086633 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6tx\" (UniqueName: \"kubernetes.io/projected/bae387b2-3007-4c94-9a6f-255bd137d7c0-kube-api-access-rw6tx\") pod \"placement-5bb67f8988-tt9dr\" (UID: \"bae387b2-3007-4c94-9a6f-255bd137d7c0\") " pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.115280 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerStarted","Data":"ba0f5fa72119c660fe2a23a0af0997d92861db2aee0bc27c7632f19e26aedea5"} Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.223841 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.435828 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.668923 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7f59c67b56-sft76" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.722315 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.888728 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5bb67f8988-tt9dr"] Feb 28 10:55:12 crc kubenswrapper[4972]: I0228 10:55:12.991471 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:55:13 crc kubenswrapper[4972]: I0228 10:55:13.024422 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:55:13 crc kubenswrapper[4972]: I0228 10:55:13.158717 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bb67f8988-tt9dr" event={"ID":"bae387b2-3007-4c94-9a6f-255bd137d7c0","Type":"ContainerStarted","Data":"f8ad0f1d2d627a04832b087477cd006b6cc6456d91a4594314b0286fa20d0024"} Feb 28 10:55:13 crc kubenswrapper[4972]: I0228 10:55:13.160432 4972 generic.go:334] "Generic (PLEG): container finished" podID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerID="75ae2c7c3c6937c7635798c50df53d6845872507f4ba65acae18046ebb3f727f" exitCode=0 Feb 28 10:55:13 crc kubenswrapper[4972]: I0228 10:55:13.162591 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerDied","Data":"75ae2c7c3c6937c7635798c50df53d6845872507f4ba65acae18046ebb3f727f"} Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.170471 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerStarted","Data":"68546d0de6a00ac3e16971daca42002df0ff2fb8fbb9edadec2f4448a28f4183"} Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.172256 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.173799 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bb67f8988-tt9dr" event={"ID":"bae387b2-3007-4c94-9a6f-255bd137d7c0","Type":"ContainerStarted","Data":"abf91a1110a66e44adc7e9590b4953d6204d31822d5c56599188c20f2a42735c"} Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.173824 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bb67f8988-tt9dr" event={"ID":"bae387b2-3007-4c94-9a6f-255bd137d7c0","Type":"ContainerStarted","Data":"3bbcb7cdb3fe6fa0fb1c28cb558d1ee3c408442a18f616112e8a2f34b05f9360"} Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.174395 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.174433 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.214545 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.190787257 podStartE2EDuration="7.214525736s" podCreationTimestamp="2026-02-28 10:55:07 +0000 UTC" firstStartedPulling="2026-02-28 10:55:08.0296664 +0000 UTC m=+1164.941674148" lastFinishedPulling="2026-02-28 10:55:13.053404899 +0000 UTC m=+1169.965412627" observedRunningTime="2026-02-28 10:55:14.206526723 +0000 UTC m=+1171.118534461" watchObservedRunningTime="2026-02-28 10:55:14.214525736 +0000 UTC m=+1171.126533474" Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.239329 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5bb67f8988-tt9dr" podStartSLOduration=3.239309218 podStartE2EDuration="3.239309218s" podCreationTimestamp="2026-02-28 10:55:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:14.229729541 +0000 UTC m=+1171.141737279" watchObservedRunningTime="2026-02-28 10:55:14.239309218 +0000 UTC m=+1171.151316946" Feb 28 10:55:14 crc kubenswrapper[4972]: I0228 10:55:14.557667 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.168:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.231281 4972 generic.go:334] "Generic (PLEG): container finished" podID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerID="3d7ff4abdbc1386075be8c3d6987ed3d6a21b6bee82abc73cef203061779a37f" exitCode=0 Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.232196 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerDied","Data":"3d7ff4abdbc1386075be8c3d6987ed3d6a21b6bee82abc73cef203061779a37f"} Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.687231 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.801849 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.801923 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.801951 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.802137 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.802207 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q89c\" (UniqueName: \"kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.802257 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts\") pod \"9efe77fe-be68-4193-a204-b5a702dbe7da\" (UID: \"9efe77fe-be68-4193-a204-b5a702dbe7da\") " Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.804387 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.810716 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts" (OuterVolumeSpecName: "scripts") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.828355 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c" (OuterVolumeSpecName: "kube-api-access-8q89c") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "kube-api-access-8q89c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.833949 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.905066 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q89c\" (UniqueName: \"kubernetes.io/projected/9efe77fe-be68-4193-a204-b5a702dbe7da-kube-api-access-8q89c\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.905099 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.905108 4972 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.905120 4972 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efe77fe-be68-4193-a204-b5a702dbe7da-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.920178 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:15 crc kubenswrapper[4972]: I0228 10:55:15.945295 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c6cf8f456-722rs" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.007587 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.025987 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data" (OuterVolumeSpecName: "config-data") pod "9efe77fe-be68-4193-a204-b5a702dbe7da" (UID: "9efe77fe-be68-4193-a204-b5a702dbe7da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.034601 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.034886 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon-log" containerID="cri-o://314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c" gracePeriod=30 Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.035024 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" containerID="cri-o://544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029" gracePeriod=30 Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.042280 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.109802 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efe77fe-be68-4193-a204-b5a702dbe7da-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.264307 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9efe77fe-be68-4193-a204-b5a702dbe7da","Type":"ContainerDied","Data":"fa7e9ac5f87b3f1424d5a787ce425d686b90ef4ba870f5ce2b8028f738746685"} Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.264386 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.264714 4972 scope.go:117] "RemoveContainer" containerID="75ae2c7c3c6937c7635798c50df53d6845872507f4ba65acae18046ebb3f727f" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.298408 4972 scope.go:117] "RemoveContainer" containerID="3d7ff4abdbc1386075be8c3d6987ed3d6a21b6bee82abc73cef203061779a37f" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.337509 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.357508 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.369628 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:16 crc kubenswrapper[4972]: E0228 10:55:16.370132 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="cinder-scheduler" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.370151 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="cinder-scheduler" Feb 28 10:55:16 crc kubenswrapper[4972]: E0228 10:55:16.370165 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="probe" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.370171 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="probe" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.370381 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="probe" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.370401 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" containerName="cinder-scheduler" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.371761 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.374869 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.381686 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.416898 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-scripts\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.416957 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.417003 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84de24a3-0eae-4ec5-af31-a515e98ef012-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.417112 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqg4m\" (UniqueName: \"kubernetes.io/projected/84de24a3-0eae-4ec5-af31-a515e98ef012-kube-api-access-mqg4m\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.417169 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.417194 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.519188 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqg4m\" (UniqueName: \"kubernetes.io/projected/84de24a3-0eae-4ec5-af31-a515e98ef012-kube-api-access-mqg4m\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.519654 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.519800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.519991 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-scripts\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.520070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.520179 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84de24a3-0eae-4ec5-af31-a515e98ef012-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.520429 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84de24a3-0eae-4ec5-af31-a515e98ef012-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.527399 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.528190 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.530016 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-scripts\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.538679 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de24a3-0eae-4ec5-af31-a515e98ef012-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.542393 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqg4m\" (UniqueName: \"kubernetes.io/projected/84de24a3-0eae-4ec5-af31-a515e98ef012-kube-api-access-mqg4m\") pod \"cinder-scheduler-0\" (UID: \"84de24a3-0eae-4ec5-af31-a515e98ef012\") " pod="openstack/cinder-scheduler-0" Feb 28 10:55:16 crc kubenswrapper[4972]: I0228 10:55:16.690033 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 28 10:55:17 crc kubenswrapper[4972]: I0228 10:55:17.326587 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 28 10:55:17 crc kubenswrapper[4972]: I0228 10:55:17.606322 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:17 crc kubenswrapper[4972]: I0228 10:55:17.776695 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 28 10:55:17 crc kubenswrapper[4972]: I0228 10:55:17.813177 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efe77fe-be68-4193-a204-b5a702dbe7da" path="/var/lib/kubelet/pods/9efe77fe-be68-4193-a204-b5a702dbe7da/volumes" Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.226753 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f94fccfd-62888" Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.334619 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84de24a3-0eae-4ec5-af31-a515e98ef012","Type":"ContainerStarted","Data":"9f71935d233d5d0cd961f23f74bc492fd87404e902f9d3320bf5ee175b9b611a"} Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.334720 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84de24a3-0eae-4ec5-af31-a515e98ef012","Type":"ContainerStarted","Data":"cc9361707cd54b196ea1c55f1595eee77d9d45c8f39726bb4278567e59f91aa5"} Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.441538 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.443218 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f59c67b56-sft76" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api-log" containerID="cri-o://cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3" gracePeriod=30 Feb 28 10:55:18 crc kubenswrapper[4972]: I0228 10:55:18.443957 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f59c67b56-sft76" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" containerID="cri-o://975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575" gracePeriod=30 Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.388891 4972 generic.go:334] "Generic (PLEG): container finished" podID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerID="586a361c55d8399ec444105b7af07b9d928fe5a427f041f28aafe2a9736e3763" exitCode=0 Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.389781 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerDied","Data":"586a361c55d8399ec444105b7af07b9d928fe5a427f041f28aafe2a9736e3763"} Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.402957 4972 generic.go:334] "Generic (PLEG): container finished" podID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerID="cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3" exitCode=143 Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.403060 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerDied","Data":"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3"} Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.439085 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.439061903 podStartE2EDuration="3.439061903s" podCreationTimestamp="2026-02-28 10:55:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:19.43139799 +0000 UTC m=+1176.343405738" watchObservedRunningTime="2026-02-28 10:55:19.439061903 +0000 UTC m=+1176.351069641" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.467186 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:50234->10.217.0.152:8443: read: connection reset by peer" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.468185 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.613731 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694415 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694515 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694629 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694717 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68w4b\" (UniqueName: \"kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694779 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694812 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.694832 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config\") pod \"9a53ec32-89b0-4473-a918-cb0d9b104caf\" (UID: \"9a53ec32-89b0-4473-a918-cb0d9b104caf\") " Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.705916 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b" (OuterVolumeSpecName: "kube-api-access-68w4b") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "kube-api-access-68w4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.730734 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.800398 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.800439 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68w4b\" (UniqueName: \"kubernetes.io/projected/9a53ec32-89b0-4473-a918-cb0d9b104caf-kube-api-access-68w4b\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.821646 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.842636 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config" (OuterVolumeSpecName: "config") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.843329 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.861621 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.880583 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9a53ec32-89b0-4473-a918-cb0d9b104caf" (UID: "9a53ec32-89b0-4473-a918-cb0d9b104caf"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.902659 4972 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.902918 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.903008 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.903067 4972 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:19 crc kubenswrapper[4972]: I0228 10:55:19.903120 4972 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a53ec32-89b0-4473-a918-cb0d9b104caf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.414750 4972 generic.go:334] "Generic (PLEG): container finished" podID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerID="544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029" exitCode=0 Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.414807 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerDied","Data":"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029"} Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.417738 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74884987d7-4szb2" event={"ID":"9a53ec32-89b0-4473-a918-cb0d9b104caf","Type":"ContainerDied","Data":"fcbc424c61716e0737789b8a7136e93c2c59bcbfd1a929a95665e9c110b95656"} Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.417788 4972 scope.go:117] "RemoveContainer" containerID="191f03e0b519f43c6debe7d7e7ead27033db34c1e41d42193d1dec9559946da1" Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.417957 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74884987d7-4szb2" Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.429318 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84de24a3-0eae-4ec5-af31-a515e98ef012","Type":"ContainerStarted","Data":"ce0065eb38dafcc64181f1836d45a265fffea6e2705fb83243fcd1d61b1b61de"} Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.496438 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.500558 4972 scope.go:117] "RemoveContainer" containerID="586a361c55d8399ec444105b7af07b9d928fe5a427f041f28aafe2a9736e3763" Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.507825 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-74884987d7-4szb2"] Feb 28 10:55:20 crc kubenswrapper[4972]: I0228 10:55:20.865029 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-66b556fdbd-nljfl" Feb 28 10:55:21 crc kubenswrapper[4972]: I0228 10:55:21.682311 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f59c67b56-sft76" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:38418->10.217.0.165:9311: read: connection reset by peer" Feb 28 10:55:21 crc kubenswrapper[4972]: I0228 10:55:21.682371 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f59c67b56-sft76" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:38426->10.217.0.165:9311: read: connection reset by peer" Feb 28 10:55:21 crc kubenswrapper[4972]: I0228 10:55:21.690889 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 28 10:55:21 crc kubenswrapper[4972]: I0228 10:55:21.806290 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" path="/var/lib/kubelet/pods/9a53ec32-89b0-4473-a918-cb0d9b104caf/volumes" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.179441 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.249966 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rcz\" (UniqueName: \"kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz\") pod \"0b2e9212-5eb8-4665-853c-37ef5f67446a\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.250041 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle\") pod \"0b2e9212-5eb8-4665-853c-37ef5f67446a\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.250204 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data\") pod \"0b2e9212-5eb8-4665-853c-37ef5f67446a\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.250287 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs\") pod \"0b2e9212-5eb8-4665-853c-37ef5f67446a\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.250318 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom\") pod \"0b2e9212-5eb8-4665-853c-37ef5f67446a\" (UID: \"0b2e9212-5eb8-4665-853c-37ef5f67446a\") " Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.250890 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs" (OuterVolumeSpecName: "logs") pod "0b2e9212-5eb8-4665-853c-37ef5f67446a" (UID: "0b2e9212-5eb8-4665-853c-37ef5f67446a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.257036 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz" (OuterVolumeSpecName: "kube-api-access-86rcz") pod "0b2e9212-5eb8-4665-853c-37ef5f67446a" (UID: "0b2e9212-5eb8-4665-853c-37ef5f67446a"). InnerVolumeSpecName "kube-api-access-86rcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.281436 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0b2e9212-5eb8-4665-853c-37ef5f67446a" (UID: "0b2e9212-5eb8-4665-853c-37ef5f67446a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.285704 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b2e9212-5eb8-4665-853c-37ef5f67446a" (UID: "0b2e9212-5eb8-4665-853c-37ef5f67446a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.312617 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data" (OuterVolumeSpecName: "config-data") pod "0b2e9212-5eb8-4665-853c-37ef5f67446a" (UID: "0b2e9212-5eb8-4665-853c-37ef5f67446a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.353560 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rcz\" (UniqueName: \"kubernetes.io/projected/0b2e9212-5eb8-4665-853c-37ef5f67446a-kube-api-access-86rcz\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.353597 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.353606 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.353616 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e9212-5eb8-4665-853c-37ef5f67446a-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.353625 4972 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b2e9212-5eb8-4665-853c-37ef5f67446a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.469441 4972 generic.go:334] "Generic (PLEG): container finished" podID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerID="975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575" exitCode=0 Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.469522 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerDied","Data":"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575"} Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.469574 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f59c67b56-sft76" event={"ID":"0b2e9212-5eb8-4665-853c-37ef5f67446a","Type":"ContainerDied","Data":"7a870468f66ed0384580066c7ba9097e41693def265b34ac871f56d9aba3620e"} Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.469607 4972 scope.go:117] "RemoveContainer" containerID="975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.469698 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f59c67b56-sft76" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.512630 4972 scope.go:117] "RemoveContainer" containerID="cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.525808 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.542045 4972 scope.go:117] "RemoveContainer" containerID="975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575" Feb 28 10:55:22 crc kubenswrapper[4972]: E0228 10:55:22.542572 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575\": container with ID starting with 975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575 not found: ID does not exist" containerID="975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.542633 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575"} err="failed to get container status \"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575\": rpc error: code = NotFound desc = could not find container \"975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575\": container with ID starting with 975c2bbb7ece112f778b16dfc7440ccdddc50152f59236228503a2e459aba575 not found: ID does not exist" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.542668 4972 scope.go:117] "RemoveContainer" containerID="cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3" Feb 28 10:55:22 crc kubenswrapper[4972]: E0228 10:55:22.543002 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3\": container with ID starting with cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3 not found: ID does not exist" containerID="cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.543050 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3"} err="failed to get container status \"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3\": rpc error: code = NotFound desc = could not find container \"cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3\": container with ID starting with cf1c09c53a51499f7e719d4e2bfdca4db87072b3bbf7402a9e7d8743d0498ef3 not found: ID does not exist" Feb 28 10:55:22 crc kubenswrapper[4972]: I0228 10:55:22.543820 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7f59c67b56-sft76"] Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.444793 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 28 10:55:23 crc kubenswrapper[4972]: E0228 10:55:23.445796 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api-log" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.445818 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api-log" Feb 28 10:55:23 crc kubenswrapper[4972]: E0228 10:55:23.445841 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-httpd" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.445851 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-httpd" Feb 28 10:55:23 crc kubenswrapper[4972]: E0228 10:55:23.445876 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.445888 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" Feb 28 10:55:23 crc kubenswrapper[4972]: E0228 10:55:23.445914 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-api" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.445939 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-api" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.446218 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-api" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.446247 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a53ec32-89b0-4473-a918-cb0d9b104caf" containerName="neutron-httpd" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.446307 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api-log" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.446331 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" containerName="barbican-api" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.447313 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.450270 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-m2drb" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.450585 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.450966 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.456864 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.589949 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.590098 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config-secret\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.590222 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.590701 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwz8q\" (UniqueName: \"kubernetes.io/projected/f069c4ff-a971-4849-9d09-4ae37dca70fc-kube-api-access-jwz8q\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.694326 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.694394 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwz8q\" (UniqueName: \"kubernetes.io/projected/f069c4ff-a971-4849-9d09-4ae37dca70fc-kube-api-access-jwz8q\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.694442 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.694540 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config-secret\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.695514 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.702295 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-openstack-config-secret\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.712756 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwz8q\" (UniqueName: \"kubernetes.io/projected/f069c4ff-a971-4849-9d09-4ae37dca70fc-kube-api-access-jwz8q\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.720903 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069c4ff-a971-4849-9d09-4ae37dca70fc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f069c4ff-a971-4849-9d09-4ae37dca70fc\") " pod="openstack/openstackclient" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.803020 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2e9212-5eb8-4665-853c-37ef5f67446a" path="/var/lib/kubelet/pods/0b2e9212-5eb8-4665-853c-37ef5f67446a/volumes" Feb 28 10:55:23 crc kubenswrapper[4972]: I0228 10:55:23.815138 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 28 10:55:24 crc kubenswrapper[4972]: I0228 10:55:24.279684 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 28 10:55:24 crc kubenswrapper[4972]: W0228 10:55:24.281223 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf069c4ff_a971_4849_9d09_4ae37dca70fc.slice/crio-ccebec98ddb87ba0f4f701fe37ae8c37f7e5aeb603fa404238038ba30fba81c5 WatchSource:0}: Error finding container ccebec98ddb87ba0f4f701fe37ae8c37f7e5aeb603fa404238038ba30fba81c5: Status 404 returned error can't find the container with id ccebec98ddb87ba0f4f701fe37ae8c37f7e5aeb603fa404238038ba30fba81c5 Feb 28 10:55:24 crc kubenswrapper[4972]: I0228 10:55:24.507981 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f069c4ff-a971-4849-9d09-4ae37dca70fc","Type":"ContainerStarted","Data":"ccebec98ddb87ba0f4f701fe37ae8c37f7e5aeb603fa404238038ba30fba81c5"} Feb 28 10:55:26 crc kubenswrapper[4972]: I0228 10:55:26.990787 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 28 10:55:28 crc kubenswrapper[4972]: I0228 10:55:28.377421 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.213568 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.214274 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-central-agent" containerID="cri-o://7925996f825b8a1aef278517beec9a3e279ad44c1f465e886cd5bcff6dfcc881" gracePeriod=30 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.214500 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-notification-agent" containerID="cri-o://64497dc29412d515d95a37308ed7726b923f50c4fd91c390e346dc327a49c600" gracePeriod=30 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.214698 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="proxy-httpd" containerID="cri-o://68546d0de6a00ac3e16971daca42002df0ff2fb8fbb9edadec2f4448a28f4183" gracePeriod=30 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.214820 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="sg-core" containerID="cri-o://ba0f5fa72119c660fe2a23a0af0997d92861db2aee0bc27c7632f19e26aedea5" gracePeriod=30 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.321824 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.170:3000/\": read tcp 10.217.0.2:34396->10.217.0.170:3000: read: connection reset by peer" Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.599585 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerID="68546d0de6a00ac3e16971daca42002df0ff2fb8fbb9edadec2f4448a28f4183" exitCode=0 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.599968 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerID="ba0f5fa72119c660fe2a23a0af0997d92861db2aee0bc27c7632f19e26aedea5" exitCode=2 Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.599634 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerDied","Data":"68546d0de6a00ac3e16971daca42002df0ff2fb8fbb9edadec2f4448a28f4183"} Feb 28 10:55:30 crc kubenswrapper[4972]: I0228 10:55:30.600014 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerDied","Data":"ba0f5fa72119c660fe2a23a0af0997d92861db2aee0bc27c7632f19e26aedea5"} Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.302549 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-98999f697-8tcfb"] Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.307315 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.309311 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.310613 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.310784 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.314540 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-98999f697-8tcfb"] Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450005 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-public-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450059 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwrcl\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-kube-api-access-lwrcl\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450592 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-etc-swift\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450718 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-log-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450754 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-combined-ca-bundle\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.450949 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-run-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.451048 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-internal-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.451133 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-config-data\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553628 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-log-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553675 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-combined-ca-bundle\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553766 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-run-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553793 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-internal-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553836 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-config-data\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553915 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-public-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.553940 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwrcl\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-kube-api-access-lwrcl\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.554000 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-etc-swift\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.554253 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-log-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.554322 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09caa13b-95cf-42c1-a35d-7c053e069909-run-httpd\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.572792 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-internal-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.572847 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-public-tls-certs\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.573430 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-combined-ca-bundle\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.573493 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09caa13b-95cf-42c1-a35d-7c053e069909-config-data\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.573994 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-etc-swift\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.576477 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwrcl\" (UniqueName: \"kubernetes.io/projected/09caa13b-95cf-42c1-a35d-7c053e069909-kube-api-access-lwrcl\") pod \"swift-proxy-98999f697-8tcfb\" (UID: \"09caa13b-95cf-42c1-a35d-7c053e069909\") " pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.616880 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerID="64497dc29412d515d95a37308ed7726b923f50c4fd91c390e346dc327a49c600" exitCode=0 Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.616920 4972 generic.go:334] "Generic (PLEG): container finished" podID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerID="7925996f825b8a1aef278517beec9a3e279ad44c1f465e886cd5bcff6dfcc881" exitCode=0 Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.616946 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerDied","Data":"64497dc29412d515d95a37308ed7726b923f50c4fd91c390e346dc327a49c600"} Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.616975 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerDied","Data":"7925996f825b8a1aef278517beec9a3e279ad44c1f465e886cd5bcff6dfcc881"} Feb 28 10:55:31 crc kubenswrapper[4972]: I0228 10:55:31.641738 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.847890 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-dmxzn"] Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.851253 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.873450 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dmxzn"] Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.933541 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-qspjr"] Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.935289 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.945242 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qspjr"] Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.985523 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29qdp\" (UniqueName: \"kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:32 crc kubenswrapper[4972]: I0228 10:55:32.985795 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.051442 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-93d4-account-create-update-w8pl2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.071605 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.074322 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.074609 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-gt8dq"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.076387 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.088673 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.088770 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.088846 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clh8q\" (UniqueName: \"kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.088939 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29qdp\" (UniqueName: \"kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.091446 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-93d4-account-create-update-w8pl2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.098929 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.112654 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gt8dq"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.119629 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29qdp\" (UniqueName: \"kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp\") pod \"nova-api-db-create-dmxzn\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.179802 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.191076 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.192294 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clh8q\" (UniqueName: \"kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.192165 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.192818 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h27kt\" (UniqueName: \"kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.192974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.193058 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94tvk\" (UniqueName: \"kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.193097 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.230040 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clh8q\" (UniqueName: \"kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q\") pod \"nova-cell0-db-create-qspjr\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.246203 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-53fe-account-create-update-b6lq2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.249717 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.252323 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.259270 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-53fe-account-create-update-b6lq2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.263856 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.295233 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h27kt\" (UniqueName: \"kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.295706 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.295768 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94tvk\" (UniqueName: \"kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.295810 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.296554 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.296892 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.322166 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h27kt\" (UniqueName: \"kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt\") pod \"nova-cell1-db-create-gt8dq\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.323568 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94tvk\" (UniqueName: \"kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk\") pod \"nova-api-93d4-account-create-update-w8pl2\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.357992 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7008-account-create-update-xtgq2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.359516 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.363095 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.368972 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7008-account-create-update-xtgq2"] Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.397847 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.398242 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7gdq\" (UniqueName: \"kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.467026 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.468419 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.500692 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.500782 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.500841 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtshc\" (UniqueName: \"kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.500889 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7gdq\" (UniqueName: \"kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.501885 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.523316 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7gdq\" (UniqueName: \"kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq\") pod \"nova-cell0-53fe-account-create-update-b6lq2\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.581395 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.603096 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.603195 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtshc\" (UniqueName: \"kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.603874 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.625162 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtshc\" (UniqueName: \"kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc\") pod \"nova-cell1-7008-account-create-update-xtgq2\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:33 crc kubenswrapper[4972]: I0228 10:55:33.738924 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.510735 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.553559 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554692 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554742 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554814 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554831 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5p77\" (UniqueName: \"kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554863 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.554929 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data\") pod \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\" (UID: \"c3144144-e9d4-4a5b-804e-2205c6f5cb00\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.555616 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.556391 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.573860 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts" (OuterVolumeSpecName: "scripts") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.573860 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77" (OuterVolumeSpecName: "kube-api-access-t5p77") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "kube-api-access-t5p77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.600524 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.659416 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.659446 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.659468 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3144144-e9d4-4a5b-804e-2205c6f5cb00-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.659477 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5p77\" (UniqueName: \"kubernetes.io/projected/c3144144-e9d4-4a5b-804e-2205c6f5cb00-kube-api-access-t5p77\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.659489 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.677752 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3144144-e9d4-4a5b-804e-2205c6f5cb00","Type":"ContainerDied","Data":"51f1fc087d80d26d52cf588d5399be26f43b07d01508a23524514db3f82a113b"} Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.677805 4972 scope.go:117] "RemoveContainer" containerID="68546d0de6a00ac3e16971daca42002df0ff2fb8fbb9edadec2f4448a28f4183" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.678054 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.682974 4972 generic.go:334] "Generic (PLEG): container finished" podID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerID="cc36b1309774da247b91f92e933c771ce7b01ef579a7b1423d5110aab679d5c7" exitCode=137 Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.683683 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerDied","Data":"cc36b1309774da247b91f92e933c771ce7b01ef579a7b1423d5110aab679d5c7"} Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.683867 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.719525 4972 scope.go:117] "RemoveContainer" containerID="ba0f5fa72119c660fe2a23a0af0997d92861db2aee0bc27c7632f19e26aedea5" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.719805 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data" (OuterVolumeSpecName: "config-data") pod "c3144144-e9d4-4a5b-804e-2205c6f5cb00" (UID: "c3144144-e9d4-4a5b-804e-2205c6f5cb00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.754025 4972 scope.go:117] "RemoveContainer" containerID="64497dc29412d515d95a37308ed7726b923f50c4fd91c390e346dc327a49c600" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.761718 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.761756 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3144144-e9d4-4a5b-804e-2205c6f5cb00-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.790101 4972 scope.go:117] "RemoveContainer" containerID="7925996f825b8a1aef278517beec9a3e279ad44c1f465e886cd5bcff6dfcc881" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.855705 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-dmxzn"] Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.911800 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967276 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967363 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967396 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967471 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967571 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967599 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4whgt\" (UniqueName: \"kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967610 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.967628 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle\") pod \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\" (UID: \"b1f9fb87-9896-48e6-9491-24b4a96ec91a\") " Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.968580 4972 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1f9fb87-9896-48e6-9491-24b4a96ec91a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.968885 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs" (OuterVolumeSpecName: "logs") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.971623 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.972947 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts" (OuterVolumeSpecName: "scripts") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:35 crc kubenswrapper[4972]: I0228 10:55:35.977019 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt" (OuterVolumeSpecName: "kube-api-access-4whgt") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "kube-api-access-4whgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.034271 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.053925 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.061248 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.071026 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4whgt\" (UniqueName: \"kubernetes.io/projected/b1f9fb87-9896-48e6-9491-24b4a96ec91a-kube-api-access-4whgt\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.071056 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.071065 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1f9fb87-9896-48e6-9491-24b4a96ec91a-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.071075 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.071086 4972 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.102521 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103025 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-central-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103044 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-central-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103064 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="sg-core" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103071 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="sg-core" Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103084 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103090 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api" Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103101 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-notification-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103107 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-notification-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103129 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api-log" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103136 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api-log" Feb 28 10:55:36 crc kubenswrapper[4972]: E0228 10:55:36.103156 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="proxy-httpd" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103162 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="proxy-httpd" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103357 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="sg-core" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103374 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103382 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-central-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103391 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" containerName="cinder-api-log" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103403 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="proxy-httpd" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103415 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" containerName="ceilometer-notification-agent" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.103592 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data" (OuterVolumeSpecName: "config-data") pod "b1f9fb87-9896-48e6-9491-24b4a96ec91a" (UID: "b1f9fb87-9896-48e6-9491-24b4a96ec91a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.105159 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.112975 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.113648 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.179553 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181445 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181511 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181572 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zrcx\" (UniqueName: \"kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181604 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181627 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181659 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181674 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.181724 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f9fb87-9896-48e6-9491-24b4a96ec91a-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.263535 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gt8dq"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288345 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288400 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288442 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288511 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288560 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288585 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.288645 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zrcx\" (UniqueName: \"kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.289007 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.289979 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.311128 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-53fe-account-create-update-b6lq2"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.320007 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.326477 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.341184 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zrcx\" (UniqueName: \"kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.360052 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.366098 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.383545 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-98999f697-8tcfb"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.508648 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7008-account-create-update-xtgq2"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.521535 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-93d4-account-create-update-w8pl2"] Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.536762 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qspjr"] Feb 28 10:55:36 crc kubenswrapper[4972]: W0228 10:55:36.545854 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d9ab92_ebdd_48e1_bcff_c7baf040dee3.slice/crio-bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb WatchSource:0}: Error finding container bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb: Status 404 returned error can't find the container with id bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.589649 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.714873 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-93d4-account-create-update-w8pl2" event={"ID":"45d9ab92-ebdd-48e1-bcff-c7baf040dee3","Type":"ContainerStarted","Data":"bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.723763 4972 generic.go:334] "Generic (PLEG): container finished" podID="8c9dd22f-f39a-48c5-9c92-3b77871d93aa" containerID="71c97183b340cff7d37bf1aab9dc2715563329ef2ba4e0e6a64ba5f365d22255" exitCode=0 Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.723847 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dmxzn" event={"ID":"8c9dd22f-f39a-48c5-9c92-3b77871d93aa","Type":"ContainerDied","Data":"71c97183b340cff7d37bf1aab9dc2715563329ef2ba4e0e6a64ba5f365d22255"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.723878 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dmxzn" event={"ID":"8c9dd22f-f39a-48c5-9c92-3b77871d93aa","Type":"ContainerStarted","Data":"900110a0f83d059eca120cc20d11eaa5ad9981f5d81c017b1291fbbf727d9799"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.726453 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qspjr" event={"ID":"e03b775f-1119-4c4b-8661-74213244355e","Type":"ContainerStarted","Data":"d9ee1c98161c3491b14f9fd1386e6364f9525895f3746c7ea00a919bd11dd650"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.730209 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gt8dq" event={"ID":"20ae66a7-72f6-445b-aa8a-f6e820d86743","Type":"ContainerStarted","Data":"b080eaf67a6d54e7254947e162be62df77ad2b581da656755e6333f33e6b5e80"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.730234 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gt8dq" event={"ID":"20ae66a7-72f6-445b-aa8a-f6e820d86743","Type":"ContainerStarted","Data":"10540143242bd695accc7a1d1ab1aaec183b8a9ba5438e14efd720b130e0171d"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.743708 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" event={"ID":"f5fd2836-500a-432f-b537-db8ddd108097","Type":"ContainerStarted","Data":"be70f4f35f581cb85aa5067da7413584a936abdde3856d046d8a38d110c47f2f"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.764600 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" event={"ID":"a679e6b9-6ad9-450f-8b9e-2d37588bfea8","Type":"ContainerStarted","Data":"01aaa1f5ceec02d6680959beec691308ab6f06f8d3eb32958dc21a8dfbd75f75"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.764642 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" event={"ID":"a679e6b9-6ad9-450f-8b9e-2d37588bfea8","Type":"ContainerStarted","Data":"4801f48193211706cac1df713c9e4e5269754c6b78b31b93c8225919bd130930"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.774957 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f069c4ff-a971-4849-9d09-4ae37dca70fc","Type":"ContainerStarted","Data":"da3944aaa7aa1c6689d5a3d4229ec7d1ccf6950f8ccd29a1c40b6290955ce178"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.781049 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-gt8dq" podStartSLOduration=3.7810134189999998 podStartE2EDuration="3.781013419s" podCreationTimestamp="2026-02-28 10:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:36.760298641 +0000 UTC m=+1193.672306369" watchObservedRunningTime="2026-02-28 10:55:36.781013419 +0000 UTC m=+1193.693021157" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.784550 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1f9fb87-9896-48e6-9491-24b4a96ec91a","Type":"ContainerDied","Data":"98550b6eac9a707fc12adae4fdd1b52357d3b67644d4d79e0e4434d030de1500"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.784608 4972 scope.go:117] "RemoveContainer" containerID="cc36b1309774da247b91f92e933c771ce7b01ef579a7b1423d5110aab679d5c7" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.784761 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.822402 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98999f697-8tcfb" event={"ID":"09caa13b-95cf-42c1-a35d-7c053e069909","Type":"ContainerStarted","Data":"3d8bd18af8ecf531dcb32bb27927ef16f8e3a15272fa3ae2efec985cad4af5c4"} Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.868448 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" podStartSLOduration=3.86842219 podStartE2EDuration="3.86842219s" podCreationTimestamp="2026-02-28 10:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:36.788068555 +0000 UTC m=+1193.700076303" watchObservedRunningTime="2026-02-28 10:55:36.86842219 +0000 UTC m=+1193.780429938" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.944533 4972 scope.go:117] "RemoveContainer" containerID="1e3eab8fad2e2ca159b4522f575dc0edaa0f66decc9afccb3fa57e3c5d2134a4" Feb 28 10:55:36 crc kubenswrapper[4972]: I0228 10:55:36.957795 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.071884221 podStartE2EDuration="13.957774955s" podCreationTimestamp="2026-02-28 10:55:23 +0000 UTC" firstStartedPulling="2026-02-28 10:55:24.292029935 +0000 UTC m=+1181.204037673" lastFinishedPulling="2026-02-28 10:55:35.177920669 +0000 UTC m=+1192.089928407" observedRunningTime="2026-02-28 10:55:36.835153851 +0000 UTC m=+1193.747161589" watchObservedRunningTime="2026-02-28 10:55:36.957774955 +0000 UTC m=+1193.869782693" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.000813 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.016870 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.038005 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.055366 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.058963 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.059050 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.058968 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.068772 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.224385 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.224958 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-logs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.224993 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94gfc\" (UniqueName: \"kubernetes.io/projected/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-kube-api-access-94gfc\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225066 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225254 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225295 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225411 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-scripts\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225476 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.225572 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.239630 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:37 crc kubenswrapper[4972]: W0228 10:55:37.270377 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34609a65_162d_48a6_bd64_1c7eff1bc3bf.slice/crio-504a73d6e20d25e3174517e55371f429eedc46625c36ee9c5f22e935f36361e7 WatchSource:0}: Error finding container 504a73d6e20d25e3174517e55371f429eedc46625c36ee9c5f22e935f36361e7: Status 404 returned error can't find the container with id 504a73d6e20d25e3174517e55371f429eedc46625c36ee9c5f22e935f36361e7 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.329768 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.329888 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.329920 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-logs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.329940 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94gfc\" (UniqueName: \"kubernetes.io/projected/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-kube-api-access-94gfc\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.329986 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330020 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330045 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330089 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-scripts\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330123 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330235 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.330549 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-logs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.338143 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.338162 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-scripts\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.338578 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.338753 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.339005 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.339868 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-config-data\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.350512 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94gfc\" (UniqueName: \"kubernetes.io/projected/8d0cdd4b-c98a-4297-b44d-9fcb336aecf0-kube-api-access-94gfc\") pod \"cinder-api-0\" (UID: \"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0\") " pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.595897 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.810930 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f9fb87-9896-48e6-9491-24b4a96ec91a" path="/var/lib/kubelet/pods/b1f9fb87-9896-48e6-9491-24b4a96ec91a/volumes" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.812345 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3144144-e9d4-4a5b-804e-2205c6f5cb00" path="/var/lib/kubelet/pods/c3144144-e9d4-4a5b-804e-2205c6f5cb00/volumes" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.835157 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98999f697-8tcfb" event={"ID":"09caa13b-95cf-42c1-a35d-7c053e069909","Type":"ContainerStarted","Data":"b1e534b14ee08a6b5d4df737a4698e66d99ba1cc11b8952a5026d5668b5929bd"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.835204 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98999f697-8tcfb" event={"ID":"09caa13b-95cf-42c1-a35d-7c053e069909","Type":"ContainerStarted","Data":"0d7ec1f2e42c0c5934fcbeed3f157dc809522830b0a652b78acddb98ce53de35"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.836797 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.836835 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.840298 4972 generic.go:334] "Generic (PLEG): container finished" podID="f5fd2836-500a-432f-b537-db8ddd108097" containerID="dfd19676c642577cd47a922d6d5a26f16344d989e82ae13af7b8f10983f41e33" exitCode=0 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.840353 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" event={"ID":"f5fd2836-500a-432f-b537-db8ddd108097","Type":"ContainerDied","Data":"dfd19676c642577cd47a922d6d5a26f16344d989e82ae13af7b8f10983f41e33"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.841903 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerStarted","Data":"504a73d6e20d25e3174517e55371f429eedc46625c36ee9c5f22e935f36361e7"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.843679 4972 generic.go:334] "Generic (PLEG): container finished" podID="45d9ab92-ebdd-48e1-bcff-c7baf040dee3" containerID="f1b696938685daeb81be165508e5ff9e604e1055bb09e9c2facc277b311948f9" exitCode=0 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.843898 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-93d4-account-create-update-w8pl2" event={"ID":"45d9ab92-ebdd-48e1-bcff-c7baf040dee3","Type":"ContainerDied","Data":"f1b696938685daeb81be165508e5ff9e604e1055bb09e9c2facc277b311948f9"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.847051 4972 generic.go:334] "Generic (PLEG): container finished" podID="a679e6b9-6ad9-450f-8b9e-2d37588bfea8" containerID="01aaa1f5ceec02d6680959beec691308ab6f06f8d3eb32958dc21a8dfbd75f75" exitCode=0 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.847185 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" event={"ID":"a679e6b9-6ad9-450f-8b9e-2d37588bfea8","Type":"ContainerDied","Data":"01aaa1f5ceec02d6680959beec691308ab6f06f8d3eb32958dc21a8dfbd75f75"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.849621 4972 generic.go:334] "Generic (PLEG): container finished" podID="e03b775f-1119-4c4b-8661-74213244355e" containerID="213c44d71e4f4791450c4c082e111096fa2e1ee5168aaffd8fa0d5410bb9118c" exitCode=0 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.849756 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qspjr" event={"ID":"e03b775f-1119-4c4b-8661-74213244355e","Type":"ContainerDied","Data":"213c44d71e4f4791450c4c082e111096fa2e1ee5168aaffd8fa0d5410bb9118c"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.851016 4972 generic.go:334] "Generic (PLEG): container finished" podID="20ae66a7-72f6-445b-aa8a-f6e820d86743" containerID="b080eaf67a6d54e7254947e162be62df77ad2b581da656755e6333f33e6b5e80" exitCode=0 Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.851197 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gt8dq" event={"ID":"20ae66a7-72f6-445b-aa8a-f6e820d86743","Type":"ContainerDied","Data":"b080eaf67a6d54e7254947e162be62df77ad2b581da656755e6333f33e6b5e80"} Feb 28 10:55:37 crc kubenswrapper[4972]: I0228 10:55:37.867207 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-98999f697-8tcfb" podStartSLOduration=6.8671861530000005 podStartE2EDuration="6.867186153s" podCreationTimestamp="2026-02-28 10:55:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:37.856403722 +0000 UTC m=+1194.768411460" watchObservedRunningTime="2026-02-28 10:55:37.867186153 +0000 UTC m=+1194.779193891" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.119855 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.172566 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69dd77478c-x64n9" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.248965 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.259057 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-564fbcbf5b-lsdwn" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-api" containerID="cri-o://e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701" gracePeriod=30 Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.259155 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-564fbcbf5b-lsdwn" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-httpd" containerID="cri-o://f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9" gracePeriod=30 Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.304871 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.377369 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c898cfcfb-6nt9v" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.491125 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts\") pod \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.491368 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29qdp\" (UniqueName: \"kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp\") pod \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\" (UID: \"8c9dd22f-f39a-48c5-9c92-3b77871d93aa\") " Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.491834 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c9dd22f-f39a-48c5-9c92-3b77871d93aa" (UID: "8c9dd22f-f39a-48c5-9c92-3b77871d93aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.493331 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.519865 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp" (OuterVolumeSpecName: "kube-api-access-29qdp") pod "8c9dd22f-f39a-48c5-9c92-3b77871d93aa" (UID: "8c9dd22f-f39a-48c5-9c92-3b77871d93aa"). InnerVolumeSpecName "kube-api-access-29qdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.595735 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29qdp\" (UniqueName: \"kubernetes.io/projected/8c9dd22f-f39a-48c5-9c92-3b77871d93aa-kube-api-access-29qdp\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.868626 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0","Type":"ContainerStarted","Data":"7356286ea96e5292e6260d088db22fd93a3b026c37838ba398623fa5cf6be981"} Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.871931 4972 generic.go:334] "Generic (PLEG): container finished" podID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerID="f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9" exitCode=0 Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.871983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerDied","Data":"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9"} Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.880525 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerStarted","Data":"444a30089d7217742ac41023b4d58164f53ae6358ae4bb2047abdaca72e305aa"} Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.880571 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerStarted","Data":"d913ab0e57f12cb89bb3636367ec7ec425cbc61f1c4d7dce03a48d1f2b6e56c6"} Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.883429 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-dmxzn" Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.891701 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-dmxzn" event={"ID":"8c9dd22f-f39a-48c5-9c92-3b77871d93aa","Type":"ContainerDied","Data":"900110a0f83d059eca120cc20d11eaa5ad9981f5d81c017b1291fbbf727d9799"} Feb 28 10:55:38 crc kubenswrapper[4972]: I0228 10:55:38.891766 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="900110a0f83d059eca120cc20d11eaa5ad9981f5d81c017b1291fbbf727d9799" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.457390 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.629441 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts\") pod \"f5fd2836-500a-432f-b537-db8ddd108097\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.630129 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtshc\" (UniqueName: \"kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc\") pod \"f5fd2836-500a-432f-b537-db8ddd108097\" (UID: \"f5fd2836-500a-432f-b537-db8ddd108097\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.639247 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc" (OuterVolumeSpecName: "kube-api-access-vtshc") pod "f5fd2836-500a-432f-b537-db8ddd108097" (UID: "f5fd2836-500a-432f-b537-db8ddd108097"). InnerVolumeSpecName "kube-api-access-vtshc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.646964 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5fd2836-500a-432f-b537-db8ddd108097" (UID: "f5fd2836-500a-432f-b537-db8ddd108097"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.732642 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5fd2836-500a-432f-b537-db8ddd108097-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.732685 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtshc\" (UniqueName: \"kubernetes.io/projected/f5fd2836-500a-432f-b537-db8ddd108097-kube-api-access-vtshc\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.775327 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.797107 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.835434 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h27kt\" (UniqueName: \"kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt\") pod \"20ae66a7-72f6-445b-aa8a-f6e820d86743\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.835591 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts\") pod \"20ae66a7-72f6-445b-aa8a-f6e820d86743\" (UID: \"20ae66a7-72f6-445b-aa8a-f6e820d86743\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.835683 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts\") pod \"e03b775f-1119-4c4b-8661-74213244355e\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.835773 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clh8q\" (UniqueName: \"kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q\") pod \"e03b775f-1119-4c4b-8661-74213244355e\" (UID: \"e03b775f-1119-4c4b-8661-74213244355e\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.839718 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q" (OuterVolumeSpecName: "kube-api-access-clh8q") pod "e03b775f-1119-4c4b-8661-74213244355e" (UID: "e03b775f-1119-4c4b-8661-74213244355e"). InnerVolumeSpecName "kube-api-access-clh8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.841752 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e03b775f-1119-4c4b-8661-74213244355e" (UID: "e03b775f-1119-4c4b-8661-74213244355e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.841767 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20ae66a7-72f6-445b-aa8a-f6e820d86743" (UID: "20ae66a7-72f6-445b-aa8a-f6e820d86743"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.843783 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt" (OuterVolumeSpecName: "kube-api-access-h27kt") pod "20ae66a7-72f6-445b-aa8a-f6e820d86743" (UID: "20ae66a7-72f6-445b-aa8a-f6e820d86743"). InnerVolumeSpecName "kube-api-access-h27kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.845165 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.847376 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.907050 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-93d4-account-create-update-w8pl2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.907116 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-93d4-account-create-update-w8pl2" event={"ID":"45d9ab92-ebdd-48e1-bcff-c7baf040dee3","Type":"ContainerDied","Data":"bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.907265 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfadc4dfd8997364556b771ff35ec15301abffc857bf11dea97d417bfa17d5bb" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.908419 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" event={"ID":"a679e6b9-6ad9-450f-8b9e-2d37588bfea8","Type":"ContainerDied","Data":"4801f48193211706cac1df713c9e4e5269754c6b78b31b93c8225919bd130930"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.908442 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4801f48193211706cac1df713c9e4e5269754c6b78b31b93c8225919bd130930" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.908574 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53fe-account-create-update-b6lq2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.912212 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gt8dq" event={"ID":"20ae66a7-72f6-445b-aa8a-f6e820d86743","Type":"ContainerDied","Data":"10540143242bd695accc7a1d1ab1aaec183b8a9ba5438e14efd720b130e0171d"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.912243 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10540143242bd695accc7a1d1ab1aaec183b8a9ba5438e14efd720b130e0171d" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.912330 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gt8dq" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.919166 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qspjr" event={"ID":"e03b775f-1119-4c4b-8661-74213244355e","Type":"ContainerDied","Data":"d9ee1c98161c3491b14f9fd1386e6364f9525895f3746c7ea00a919bd11dd650"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.919208 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9ee1c98161c3491b14f9fd1386e6364f9525895f3746c7ea00a919bd11dd650" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.919266 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qspjr" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.921754 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0","Type":"ContainerStarted","Data":"cb35d77ee9df2c4de393c0b11c7a03bfb0e3a3b952a03ad6d747919545aacc0d"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.923325 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.923784 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7008-account-create-update-xtgq2" event={"ID":"f5fd2836-500a-432f-b537-db8ddd108097","Type":"ContainerDied","Data":"be70f4f35f581cb85aa5067da7413584a936abdde3856d046d8a38d110c47f2f"} Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.923802 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be70f4f35f581cb85aa5067da7413584a936abdde3856d046d8a38d110c47f2f" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.938690 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts\") pod \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.939683 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94tvk\" (UniqueName: \"kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk\") pod \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\" (UID: \"45d9ab92-ebdd-48e1-bcff-c7baf040dee3\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.939869 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts\") pod \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.940063 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7gdq\" (UniqueName: \"kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq\") pod \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\" (UID: \"a679e6b9-6ad9-450f-8b9e-2d37588bfea8\") " Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.940735 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h27kt\" (UniqueName: \"kubernetes.io/projected/20ae66a7-72f6-445b-aa8a-f6e820d86743-kube-api-access-h27kt\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.940813 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ae66a7-72f6-445b-aa8a-f6e820d86743-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.940918 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e03b775f-1119-4c4b-8661-74213244355e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.941009 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clh8q\" (UniqueName: \"kubernetes.io/projected/e03b775f-1119-4c4b-8661-74213244355e-kube-api-access-clh8q\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.942251 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45d9ab92-ebdd-48e1-bcff-c7baf040dee3" (UID: "45d9ab92-ebdd-48e1-bcff-c7baf040dee3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.943497 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a679e6b9-6ad9-450f-8b9e-2d37588bfea8" (UID: "a679e6b9-6ad9-450f-8b9e-2d37588bfea8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.947998 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq" (OuterVolumeSpecName: "kube-api-access-p7gdq") pod "a679e6b9-6ad9-450f-8b9e-2d37588bfea8" (UID: "a679e6b9-6ad9-450f-8b9e-2d37588bfea8"). InnerVolumeSpecName "kube-api-access-p7gdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:39 crc kubenswrapper[4972]: I0228 10:55:39.949271 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk" (OuterVolumeSpecName: "kube-api-access-94tvk") pod "45d9ab92-ebdd-48e1-bcff-c7baf040dee3" (UID: "45d9ab92-ebdd-48e1-bcff-c7baf040dee3"). InnerVolumeSpecName "kube-api-access-94tvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.044091 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.044131 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7gdq\" (UniqueName: \"kubernetes.io/projected/a679e6b9-6ad9-450f-8b9e-2d37588bfea8-kube-api-access-p7gdq\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.044145 4972 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.044155 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94tvk\" (UniqueName: \"kubernetes.io/projected/45d9ab92-ebdd-48e1-bcff-c7baf040dee3-kube-api-access-94tvk\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.937637 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerStarted","Data":"06d1307f3df94342757d9c426e925decd033a3d6ecc197c4ff93d496c14a246f"} Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.939830 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d0cdd4b-c98a-4297-b44d-9fcb336aecf0","Type":"ContainerStarted","Data":"ee5bf0b7e3dddeb58e425aa49fe630faf3768e0657e41ffe8ccff5ea2c95ed32"} Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.940096 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 28 10:55:40 crc kubenswrapper[4972]: I0228 10:55:40.963980 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.963952507 podStartE2EDuration="4.963952507s" podCreationTimestamp="2026-02-28 10:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:55:40.956769827 +0000 UTC m=+1197.868777565" watchObservedRunningTime="2026-02-28 10:55:40.963952507 +0000 UTC m=+1197.875960245" Feb 28 10:55:41 crc kubenswrapper[4972]: I0228 10:55:41.181429 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:41 crc kubenswrapper[4972]: I0228 10:55:41.661083 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964000 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerStarted","Data":"52429e836abbb473ef137c36b98696dceaaa506139d89385e8dd51c6f7e3e3b2"} Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964186 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-central-agent" containerID="cri-o://d913ab0e57f12cb89bb3636367ec7ec425cbc61f1c4d7dce03a48d1f2b6e56c6" gracePeriod=30 Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964509 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="proxy-httpd" containerID="cri-o://52429e836abbb473ef137c36b98696dceaaa506139d89385e8dd51c6f7e3e3b2" gracePeriod=30 Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964571 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-notification-agent" containerID="cri-o://444a30089d7217742ac41023b4d58164f53ae6358ae4bb2047abdaca72e305aa" gracePeriod=30 Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964626 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="sg-core" containerID="cri-o://06d1307f3df94342757d9c426e925decd033a3d6ecc197c4ff93d496c14a246f" gracePeriod=30 Feb 28 10:55:42 crc kubenswrapper[4972]: I0228 10:55:42.964740 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.010453 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.282403749 podStartE2EDuration="7.01042926s" podCreationTimestamp="2026-02-28 10:55:36 +0000 UTC" firstStartedPulling="2026-02-28 10:55:37.273903274 +0000 UTC m=+1194.185911012" lastFinishedPulling="2026-02-28 10:55:42.001928775 +0000 UTC m=+1198.913936523" observedRunningTime="2026-02-28 10:55:42.997445437 +0000 UTC m=+1199.909453175" watchObservedRunningTime="2026-02-28 10:55:43.01042926 +0000 UTC m=+1199.922436998" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.270295 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.402798 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5bb67f8988-tt9dr" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.470846 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.471749 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5cf557df54-q4jtg" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-log" containerID="cri-o://d19859f8e5fd6ed66512dfff8d2897b1d9e4f19dd48408278508fc8015a42b39" gracePeriod=30 Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.473793 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5cf557df54-q4jtg" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-api" containerID="cri-o://b37bfd33de2192e1c15aec2849e6a120e2737b171938309f23f7e3fbad3ea6bc" gracePeriod=30 Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.585591 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-56fjs"] Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586093 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a679e6b9-6ad9-450f-8b9e-2d37588bfea8" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586111 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a679e6b9-6ad9-450f-8b9e-2d37588bfea8" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586127 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ae66a7-72f6-445b-aa8a-f6e820d86743" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586135 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ae66a7-72f6-445b-aa8a-f6e820d86743" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586149 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5fd2836-500a-432f-b537-db8ddd108097" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586155 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5fd2836-500a-432f-b537-db8ddd108097" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586164 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e03b775f-1119-4c4b-8661-74213244355e" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586170 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e03b775f-1119-4c4b-8661-74213244355e" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586181 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d9ab92-ebdd-48e1-bcff-c7baf040dee3" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586189 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d9ab92-ebdd-48e1-bcff-c7baf040dee3" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: E0228 10:55:43.586225 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9dd22f-f39a-48c5-9c92-3b77871d93aa" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586233 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9dd22f-f39a-48c5-9c92-3b77871d93aa" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586398 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9dd22f-f39a-48c5-9c92-3b77871d93aa" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586421 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ae66a7-72f6-445b-aa8a-f6e820d86743" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586435 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e03b775f-1119-4c4b-8661-74213244355e" containerName="mariadb-database-create" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586443 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5fd2836-500a-432f-b537-db8ddd108097" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586458 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d9ab92-ebdd-48e1-bcff-c7baf040dee3" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.586485 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a679e6b9-6ad9-450f-8b9e-2d37588bfea8" containerName="mariadb-account-create-update" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.587115 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.604933 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-56fjs"] Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.605272 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.605620 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.605842 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-25pgq" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.726444 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.726647 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.726671 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.726746 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xm8\" (UniqueName: \"kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.829566 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.829626 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.829657 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.829741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xm8\" (UniqueName: \"kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.840023 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.840547 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.841848 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.847015 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.847127 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:43 crc kubenswrapper[4972]: I0228 10:55:43.850915 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xm8\" (UniqueName: \"kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8\") pod \"nova-cell0-conductor-db-sync-56fjs\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.033611 4972 generic.go:334] "Generic (PLEG): container finished" podID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerID="52429e836abbb473ef137c36b98696dceaaa506139d89385e8dd51c6f7e3e3b2" exitCode=0 Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.037896 4972 generic.go:334] "Generic (PLEG): container finished" podID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerID="06d1307f3df94342757d9c426e925decd033a3d6ecc197c4ff93d496c14a246f" exitCode=2 Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.037939 4972 generic.go:334] "Generic (PLEG): container finished" podID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerID="444a30089d7217742ac41023b4d58164f53ae6358ae4bb2047abdaca72e305aa" exitCode=0 Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.037956 4972 generic.go:334] "Generic (PLEG): container finished" podID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerID="d913ab0e57f12cb89bb3636367ec7ec425cbc61f1c4d7dce03a48d1f2b6e56c6" exitCode=0 Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.034031 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerDied","Data":"52429e836abbb473ef137c36b98696dceaaa506139d89385e8dd51c6f7e3e3b2"} Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.038577 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerDied","Data":"06d1307f3df94342757d9c426e925decd033a3d6ecc197c4ff93d496c14a246f"} Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.038606 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerDied","Data":"444a30089d7217742ac41023b4d58164f53ae6358ae4bb2047abdaca72e305aa"} Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.038618 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerDied","Data":"d913ab0e57f12cb89bb3636367ec7ec425cbc61f1c4d7dce03a48d1f2b6e56c6"} Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.064765 4972 generic.go:334] "Generic (PLEG): container finished" podID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerID="d19859f8e5fd6ed66512dfff8d2897b1d9e4f19dd48408278508fc8015a42b39" exitCode=143 Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.065309 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerDied","Data":"d19859f8e5fd6ed66512dfff8d2897b1d9e4f19dd48408278508fc8015a42b39"} Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.118514 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-25pgq" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.126244 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.268990 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.394482 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.394886 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.394965 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.395015 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.395120 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.395223 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zrcx\" (UniqueName: \"kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.395292 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts\") pod \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\" (UID: \"34609a65-162d-48a6-bd64-1c7eff1bc3bf\") " Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.397027 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.397565 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.414743 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx" (OuterVolumeSpecName: "kube-api-access-5zrcx") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "kube-api-access-5zrcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.476689 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts" (OuterVolumeSpecName: "scripts") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.498052 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zrcx\" (UniqueName: \"kubernetes.io/projected/34609a65-162d-48a6-bd64-1c7eff1bc3bf-kube-api-access-5zrcx\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.498099 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.498112 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.498124 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34609a65-162d-48a6-bd64-1c7eff1bc3bf-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.518737 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.601114 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.608408 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data" (OuterVolumeSpecName: "config-data") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.633481 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34609a65-162d-48a6-bd64-1c7eff1bc3bf" (UID: "34609a65-162d-48a6-bd64-1c7eff1bc3bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.703917 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.703957 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34609a65-162d-48a6-bd64-1c7eff1bc3bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.813623 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-56fjs"] Feb 28 10:55:44 crc kubenswrapper[4972]: I0228 10:55:44.882850 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.009701 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config\") pod \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.009898 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config\") pod \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.010029 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs\") pod \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.010068 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tw5w\" (UniqueName: \"kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w\") pod \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.010176 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle\") pod \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\" (UID: \"eab6dc5d-2bfc-4dae-9309-e4e911223c2a\") " Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.015227 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w" (OuterVolumeSpecName: "kube-api-access-6tw5w") pod "eab6dc5d-2bfc-4dae-9309-e4e911223c2a" (UID: "eab6dc5d-2bfc-4dae-9309-e4e911223c2a"). InnerVolumeSpecName "kube-api-access-6tw5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.016960 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "eab6dc5d-2bfc-4dae-9309-e4e911223c2a" (UID: "eab6dc5d-2bfc-4dae-9309-e4e911223c2a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.077973 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-56fjs" event={"ID":"6e7caa58-92af-435f-9d46-6f63ed19a6e6","Type":"ContainerStarted","Data":"04fd827ecb6a823ba4729f69fafc691bbbde95e40daeea43d53765382569b6b2"} Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.078833 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eab6dc5d-2bfc-4dae-9309-e4e911223c2a" (UID: "eab6dc5d-2bfc-4dae-9309-e4e911223c2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.081256 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34609a65-162d-48a6-bd64-1c7eff1bc3bf","Type":"ContainerDied","Data":"504a73d6e20d25e3174517e55371f429eedc46625c36ee9c5f22e935f36361e7"} Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.081326 4972 scope.go:117] "RemoveContainer" containerID="52429e836abbb473ef137c36b98696dceaaa506139d89385e8dd51c6f7e3e3b2" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.081514 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.089616 4972 generic.go:334] "Generic (PLEG): container finished" podID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerID="e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701" exitCode=0 Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.089672 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerDied","Data":"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701"} Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.089703 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564fbcbf5b-lsdwn" event={"ID":"eab6dc5d-2bfc-4dae-9309-e4e911223c2a","Type":"ContainerDied","Data":"4f89459f4579cc3177b12241f938868587e3538a6c0b005bb0c8f66f8dfce383"} Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.089776 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564fbcbf5b-lsdwn" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.090684 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config" (OuterVolumeSpecName: "config") pod "eab6dc5d-2bfc-4dae-9309-e4e911223c2a" (UID: "eab6dc5d-2bfc-4dae-9309-e4e911223c2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.112294 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.112320 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.112330 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.112340 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tw5w\" (UniqueName: \"kubernetes.io/projected/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-kube-api-access-6tw5w\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.117244 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "eab6dc5d-2bfc-4dae-9309-e4e911223c2a" (UID: "eab6dc5d-2bfc-4dae-9309-e4e911223c2a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.204452 4972 scope.go:117] "RemoveContainer" containerID="06d1307f3df94342757d9c426e925decd033a3d6ecc197c4ff93d496c14a246f" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.214182 4972 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab6dc5d-2bfc-4dae-9309-e4e911223c2a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.215241 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.227056 4972 scope.go:117] "RemoveContainer" containerID="444a30089d7217742ac41023b4d58164f53ae6358ae4bb2047abdaca72e305aa" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.231679 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.251385 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252805 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-central-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252830 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-central-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252855 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-api" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252862 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-api" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252879 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-notification-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252886 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-notification-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252900 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="sg-core" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252908 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="sg-core" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252927 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="proxy-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252934 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="proxy-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.252944 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.252950 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253539 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="sg-core" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253567 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-api" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253586 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" containerName="neutron-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253594 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="proxy-httpd" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253601 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-notification-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.253610 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" containerName="ceilometer-central-agent" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.256346 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.261513 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.261588 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.261668 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318442 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318527 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df2cm\" (UniqueName: \"kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318552 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318591 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318696 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318713 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.318756 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.339771 4972 scope.go:117] "RemoveContainer" containerID="d913ab0e57f12cb89bb3636367ec7ec425cbc61f1c4d7dce03a48d1f2b6e56c6" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.372659 4972 scope.go:117] "RemoveContainer" containerID="f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.403529 4972 scope.go:117] "RemoveContainer" containerID="e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421117 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421168 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421223 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421297 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421325 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df2cm\" (UniqueName: \"kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421354 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.421414 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.422687 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.423528 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.439387 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.442943 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.443010 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.443293 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.443717 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df2cm\" (UniqueName: \"kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.445435 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data\") pod \"ceilometer-0\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.451646 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-564fbcbf5b-lsdwn"] Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.458473 4972 scope.go:117] "RemoveContainer" containerID="f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.460322 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9\": container with ID starting with f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9 not found: ID does not exist" containerID="f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.460380 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9"} err="failed to get container status \"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9\": rpc error: code = NotFound desc = could not find container \"f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9\": container with ID starting with f421b4087e5b3726968b1f022a265a23d807c054897a0de8aa279f97af6454a9 not found: ID does not exist" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.460404 4972 scope.go:117] "RemoveContainer" containerID="e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701" Feb 28 10:55:45 crc kubenswrapper[4972]: E0228 10:55:45.461036 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701\": container with ID starting with e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701 not found: ID does not exist" containerID="e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.461114 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701"} err="failed to get container status \"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701\": rpc error: code = NotFound desc = could not find container \"e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701\": container with ID starting with e78a2df1f888d0bdb2962cfd3dd8062873c249e4a81cb1ead85289048537d701 not found: ID does not exist" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.595401 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.805811 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34609a65-162d-48a6-bd64-1c7eff1bc3bf" path="/var/lib/kubelet/pods/34609a65-162d-48a6-bd64-1c7eff1bc3bf/volumes" Feb 28 10:55:45 crc kubenswrapper[4972]: I0228 10:55:45.808976 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab6dc5d-2bfc-4dae-9309-e4e911223c2a" path="/var/lib/kubelet/pods/eab6dc5d-2bfc-4dae-9309-e4e911223c2a/volumes" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.108190 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:46 crc kubenswrapper[4972]: W0228 10:55:46.144377 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f00eb70_1083_4ac2_bd6d_9c9d17fae7f2.slice/crio-a0621e73e5c0f7478ccace06c746651ae6821f60a14d4343ffe70eb7b30fb022 WatchSource:0}: Error finding container a0621e73e5c0f7478ccace06c746651ae6821f60a14d4343ffe70eb7b30fb022: Status 404 returned error can't find the container with id a0621e73e5c0f7478ccace06c746651ae6821f60a14d4343ffe70eb7b30fb022 Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.529987 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.652648 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.652736 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.652823 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.652853 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.652914 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.653032 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.653082 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmvvq\" (UniqueName: \"kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq\") pod \"71b9423f-92b6-4085-b104-4a13a12d7af8\" (UID: \"71b9423f-92b6-4085-b104-4a13a12d7af8\") " Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.658210 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs" (OuterVolumeSpecName: "logs") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.658751 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71b9423f-92b6-4085-b104-4a13a12d7af8-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.659016 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-98999f697-8tcfb" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.667093 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq" (OuterVolumeSpecName: "kube-api-access-vmvvq") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "kube-api-access-vmvvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.712976 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.740296 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.742076 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data" (OuterVolumeSpecName: "config-data") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.760887 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmvvq\" (UniqueName: \"kubernetes.io/projected/71b9423f-92b6-4085-b104-4a13a12d7af8-kube-api-access-vmvvq\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.760926 4972 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.760937 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.760948 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.875288 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts" (OuterVolumeSpecName: "scripts") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.920278 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "71b9423f-92b6-4085-b104-4a13a12d7af8" (UID: "71b9423f-92b6-4085-b104-4a13a12d7af8"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.966485 4972 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/71b9423f-92b6-4085-b104-4a13a12d7af8-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:46 crc kubenswrapper[4972]: I0228 10:55:46.966539 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b9423f-92b6-4085-b104-4a13a12d7af8-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.139349 4972 generic.go:334] "Generic (PLEG): container finished" podID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerID="b37bfd33de2192e1c15aec2849e6a120e2737b171938309f23f7e3fbad3ea6bc" exitCode=0 Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.139836 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerDied","Data":"b37bfd33de2192e1c15aec2849e6a120e2737b171938309f23f7e3fbad3ea6bc"} Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.141793 4972 generic.go:334] "Generic (PLEG): container finished" podID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerID="314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c" exitCode=137 Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.141854 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerDied","Data":"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c"} Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.141911 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c898cfcfb-6nt9v" event={"ID":"71b9423f-92b6-4085-b104-4a13a12d7af8","Type":"ContainerDied","Data":"e5ce21101515d7be8f05d51d461295627f4f68e109e8ea06b4e7503a2e605591"} Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.141932 4972 scope.go:117] "RemoveContainer" containerID="544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.142120 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c898cfcfb-6nt9v" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.146538 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerStarted","Data":"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd"} Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.146588 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerStarted","Data":"a0621e73e5c0f7478ccace06c746651ae6821f60a14d4343ffe70eb7b30fb022"} Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.202484 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.217239 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c898cfcfb-6nt9v"] Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.373805 4972 scope.go:117] "RemoveContainer" containerID="314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.524200 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.536985 4972 scope.go:117] "RemoveContainer" containerID="544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029" Feb 28 10:55:47 crc kubenswrapper[4972]: E0228 10:55:47.537536 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029\": container with ID starting with 544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029 not found: ID does not exist" containerID="544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.537586 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029"} err="failed to get container status \"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029\": rpc error: code = NotFound desc = could not find container \"544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029\": container with ID starting with 544a37e1b15c878e81990a4ca596c03b3611086f68e1fa3d84cf93fe6dea1029 not found: ID does not exist" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.537616 4972 scope.go:117] "RemoveContainer" containerID="314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c" Feb 28 10:55:47 crc kubenswrapper[4972]: E0228 10:55:47.537898 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c\": container with ID starting with 314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c not found: ID does not exist" containerID="314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.537931 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c"} err="failed to get container status \"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c\": rpc error: code = NotFound desc = could not find container \"314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c\": container with ID starting with 314c71278a1eb92a7917f985d4a38528466e3cee38d8ba94b07bf89ac655328c not found: ID does not exist" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581124 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581252 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581398 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581428 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581493 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmn6j\" (UniqueName: \"kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581539 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.581598 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs\") pod \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\" (UID: \"531e2283-6b74-430f-a2ce-9f1a0e62f11f\") " Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.583573 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs" (OuterVolumeSpecName: "logs") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.598922 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j" (OuterVolumeSpecName: "kube-api-access-vmn6j") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "kube-api-access-vmn6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.599044 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts" (OuterVolumeSpecName: "scripts") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.670226 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data" (OuterVolumeSpecName: "config-data") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.684414 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.684473 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.684487 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/531e2283-6b74-430f-a2ce-9f1a0e62f11f-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.684501 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmn6j\" (UniqueName: \"kubernetes.io/projected/531e2283-6b74-430f-a2ce-9f1a0e62f11f-kube-api-access-vmn6j\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.767662 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.790553 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.809384 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.815686 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" path="/var/lib/kubelet/pods/71b9423f-92b6-4085-b104-4a13a12d7af8/volumes" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.826359 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "531e2283-6b74-430f-a2ce-9f1a0e62f11f" (UID: "531e2283-6b74-430f-a2ce-9f1a0e62f11f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.892905 4972 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:47 crc kubenswrapper[4972]: I0228 10:55:47.892960 4972 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/531e2283-6b74-430f-a2ce-9f1a0e62f11f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.166120 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cf557df54-q4jtg" event={"ID":"531e2283-6b74-430f-a2ce-9f1a0e62f11f","Type":"ContainerDied","Data":"0e7f82e49b519140b09300846b7a4de854aaea39b1cf5b67bf606b35b6d88c2b"} Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.166122 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cf557df54-q4jtg" Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.166183 4972 scope.go:117] "RemoveContainer" containerID="b37bfd33de2192e1c15aec2849e6a120e2737b171938309f23f7e3fbad3ea6bc" Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.171429 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerStarted","Data":"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a"} Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.227267 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.236588 4972 scope.go:117] "RemoveContainer" containerID="d19859f8e5fd6ed66512dfff8d2897b1d9e4f19dd48408278508fc8015a42b39" Feb 28 10:55:48 crc kubenswrapper[4972]: I0228 10:55:48.237725 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5cf557df54-q4jtg"] Feb 28 10:55:49 crc kubenswrapper[4972]: I0228 10:55:49.197492 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerStarted","Data":"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26"} Feb 28 10:55:49 crc kubenswrapper[4972]: I0228 10:55:49.767081 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 28 10:55:49 crc kubenswrapper[4972]: I0228 10:55:49.803614 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" path="/var/lib/kubelet/pods/531e2283-6b74-430f-a2ce-9f1a0e62f11f/volumes" Feb 28 10:55:51 crc kubenswrapper[4972]: I0228 10:55:51.433180 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.290234 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerStarted","Data":"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610"} Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.290667 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-central-agent" containerID="cri-o://75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd" gracePeriod=30 Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.291023 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.291018 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="proxy-httpd" containerID="cri-o://f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610" gracePeriod=30 Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.291045 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="sg-core" containerID="cri-o://5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26" gracePeriod=30 Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.291061 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-notification-agent" containerID="cri-o://ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a" gracePeriod=30 Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.303603 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-56fjs" event={"ID":"6e7caa58-92af-435f-9d46-6f63ed19a6e6","Type":"ContainerStarted","Data":"a6cdca71b0d7d252124e9d485c53110feb4235f37dbb35cf332a1d2a62d6750b"} Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.321026 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.832051457 podStartE2EDuration="11.321006589s" podCreationTimestamp="2026-02-28 10:55:45 +0000 UTC" firstStartedPulling="2026-02-28 10:55:46.1475327 +0000 UTC m=+1203.059540438" lastFinishedPulling="2026-02-28 10:55:55.636487832 +0000 UTC m=+1212.548495570" observedRunningTime="2026-02-28 10:55:56.311223276 +0000 UTC m=+1213.223231014" watchObservedRunningTime="2026-02-28 10:55:56.321006589 +0000 UTC m=+1213.233014327" Feb 28 10:55:56 crc kubenswrapper[4972]: I0228 10:55:56.344138 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-56fjs" podStartSLOduration=2.530839017 podStartE2EDuration="13.344121804s" podCreationTimestamp="2026-02-28 10:55:43 +0000 UTC" firstStartedPulling="2026-02-28 10:55:44.82373368 +0000 UTC m=+1201.735741418" lastFinishedPulling="2026-02-28 10:55:55.637016467 +0000 UTC m=+1212.549024205" observedRunningTime="2026-02-28 10:55:56.339702871 +0000 UTC m=+1213.251710609" watchObservedRunningTime="2026-02-28 10:55:56.344121804 +0000 UTC m=+1213.256129542" Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.320666 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerID="f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610" exitCode=0 Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.321083 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerID="5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26" exitCode=2 Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.321096 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerID="ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a" exitCode=0 Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.320847 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerDied","Data":"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610"} Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.321217 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerDied","Data":"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26"} Feb 28 10:55:57 crc kubenswrapper[4972]: I0228 10:55:57.321250 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerDied","Data":"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a"} Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.096576 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142160 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142393 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df2cm\" (UniqueName: \"kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142434 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142518 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142544 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142561 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.142665 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data\") pod \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\" (UID: \"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2\") " Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.143401 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.143748 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.154733 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts" (OuterVolumeSpecName: "scripts") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.155302 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm" (OuterVolumeSpecName: "kube-api-access-df2cm") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "kube-api-access-df2cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.198218 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.217995 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251481 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251521 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df2cm\" (UniqueName: \"kubernetes.io/projected/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-kube-api-access-df2cm\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251536 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251547 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251561 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.251572 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.256044 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data" (OuterVolumeSpecName: "config-data") pod "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" (UID: "9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.345019 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerID="75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd" exitCode=0 Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.345101 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerDied","Data":"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd"} Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.345136 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2","Type":"ContainerDied","Data":"a0621e73e5c0f7478ccace06c746651ae6821f60a14d4343ffe70eb7b30fb022"} Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.345102 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.345185 4972 scope.go:117] "RemoveContainer" containerID="f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.353565 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.377156 4972 scope.go:117] "RemoveContainer" containerID="5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.380587 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.390956 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.403789 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404230 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404249 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404263 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="sg-core" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404270 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="sg-core" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404277 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="proxy-httpd" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404283 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="proxy-httpd" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404297 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-central-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404303 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-central-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404320 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-notification-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404326 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-notification-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404347 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-api" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404355 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-api" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404367 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-log" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404374 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-log" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.404385 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon-log" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404391 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon-log" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404600 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-notification-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404623 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="ceilometer-central-agent" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404641 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-api" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404657 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="531e2283-6b74-430f-a2ce-9f1a0e62f11f" containerName="placement-log" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404671 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404681 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="sg-core" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404696 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b9423f-92b6-4085-b104-4a13a12d7af8" containerName="horizon-log" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.404709 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" containerName="proxy-httpd" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.408068 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.411074 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.411327 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.415521 4972 scope.go:117] "RemoveContainer" containerID="ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.419607 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.451361 4972 scope.go:117] "RemoveContainer" containerID="75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455470 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455554 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455591 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455657 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455693 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455763 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b7wx\" (UniqueName: \"kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.455809 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.473664 4972 scope.go:117] "RemoveContainer" containerID="f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.474238 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610\": container with ID starting with f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610 not found: ID does not exist" containerID="f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.474337 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610"} err="failed to get container status \"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610\": rpc error: code = NotFound desc = could not find container \"f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610\": container with ID starting with f56ad7e1b307e340cb998137094960dc6d13f30a72bb9f294083586f4da65610 not found: ID does not exist" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.474420 4972 scope.go:117] "RemoveContainer" containerID="5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.474825 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26\": container with ID starting with 5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26 not found: ID does not exist" containerID="5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.474901 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26"} err="failed to get container status \"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26\": rpc error: code = NotFound desc = could not find container \"5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26\": container with ID starting with 5c469be0927c28b24c497f0ccbfa7c3878dfdccff5bb7379b6add5608280fa26 not found: ID does not exist" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.474961 4972 scope.go:117] "RemoveContainer" containerID="ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.475304 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a\": container with ID starting with ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a not found: ID does not exist" containerID="ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.475379 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a"} err="failed to get container status \"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a\": rpc error: code = NotFound desc = could not find container \"ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a\": container with ID starting with ee7122402e0582628008b9e7581476b8bcba250489ac58b62d7866a44243fb2a not found: ID does not exist" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.475443 4972 scope.go:117] "RemoveContainer" containerID="75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd" Feb 28 10:55:59 crc kubenswrapper[4972]: E0228 10:55:59.475973 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd\": container with ID starting with 75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd not found: ID does not exist" containerID="75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.476032 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd"} err="failed to get container status \"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd\": rpc error: code = NotFound desc = could not find container \"75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd\": container with ID starting with 75c949962c7cd630a5a0b54dadcdddaba71979ec23e546757bb32c6badcf94dd not found: ID does not exist" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558287 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b7wx\" (UniqueName: \"kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558355 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558400 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558422 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558516 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.558550 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.559419 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.559530 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.562615 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.563261 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.563532 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.566296 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.581296 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b7wx\" (UniqueName: \"kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx\") pod \"ceilometer-0\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.741965 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:55:59 crc kubenswrapper[4972]: I0228 10:55:59.808013 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2" path="/var/lib/kubelet/pods/9f00eb70-1083-4ac2-bd6d-9c9d17fae7f2/volumes" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.153292 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537936-lwr9x"] Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.155293 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.157546 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.159869 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.160155 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.172865 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537936-lwr9x"] Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.229910 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.294702 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zv5\" (UniqueName: \"kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5\") pod \"auto-csr-approver-29537936-lwr9x\" (UID: \"bcceffbe-fb98-48de-a29d-51bd75e00c7a\") " pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.356181 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerStarted","Data":"418c286642128f3b7fa56e8049db8e35775f2f8fb5d0b4622e2bf1059a455c9f"} Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.396998 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zv5\" (UniqueName: \"kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5\") pod \"auto-csr-approver-29537936-lwr9x\" (UID: \"bcceffbe-fb98-48de-a29d-51bd75e00c7a\") " pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.417130 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zv5\" (UniqueName: \"kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5\") pod \"auto-csr-approver-29537936-lwr9x\" (UID: \"bcceffbe-fb98-48de-a29d-51bd75e00c7a\") " pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.479808 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:00 crc kubenswrapper[4972]: I0228 10:56:00.942356 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537936-lwr9x"] Feb 28 10:56:00 crc kubenswrapper[4972]: W0228 10:56:00.952813 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcceffbe_fb98_48de_a29d_51bd75e00c7a.slice/crio-76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9 WatchSource:0}: Error finding container 76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9: Status 404 returned error can't find the container with id 76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9 Feb 28 10:56:01 crc kubenswrapper[4972]: I0228 10:56:01.368585 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" event={"ID":"bcceffbe-fb98-48de-a29d-51bd75e00c7a","Type":"ContainerStarted","Data":"76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9"} Feb 28 10:56:01 crc kubenswrapper[4972]: I0228 10:56:01.371414 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerStarted","Data":"203b4deaec4bdfa1707b66a279e1b87d04e1077fa797daafd3535b78e039fa7b"} Feb 28 10:56:02 crc kubenswrapper[4972]: I0228 10:56:02.385357 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerStarted","Data":"2acd85bd4eb3d7b200e659fd3c24122828d037ed5fe7420bd3f2b4926ff7537c"} Feb 28 10:56:02 crc kubenswrapper[4972]: I0228 10:56:02.389290 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" event={"ID":"bcceffbe-fb98-48de-a29d-51bd75e00c7a","Type":"ContainerStarted","Data":"ea4de5fe653a69d74be9cba96ff851fee729779e9a8925a71abdc1af2532ebd2"} Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.403308 4972 generic.go:334] "Generic (PLEG): container finished" podID="bcceffbe-fb98-48de-a29d-51bd75e00c7a" containerID="ea4de5fe653a69d74be9cba96ff851fee729779e9a8925a71abdc1af2532ebd2" exitCode=0 Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.403789 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" event={"ID":"bcceffbe-fb98-48de-a29d-51bd75e00c7a","Type":"ContainerDied","Data":"ea4de5fe653a69d74be9cba96ff851fee729779e9a8925a71abdc1af2532ebd2"} Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.407512 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerStarted","Data":"8c66ad1568a2ef2d47315e2a1a2da762c4275d7a47c5f80598667d01eb85a82f"} Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.760784 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.873557 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8zv5\" (UniqueName: \"kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5\") pod \"bcceffbe-fb98-48de-a29d-51bd75e00c7a\" (UID: \"bcceffbe-fb98-48de-a29d-51bd75e00c7a\") " Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.888099 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5" (OuterVolumeSpecName: "kube-api-access-s8zv5") pod "bcceffbe-fb98-48de-a29d-51bd75e00c7a" (UID: "bcceffbe-fb98-48de-a29d-51bd75e00c7a"). InnerVolumeSpecName "kube-api-access-s8zv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:03 crc kubenswrapper[4972]: I0228 10:56:03.976517 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8zv5\" (UniqueName: \"kubernetes.io/projected/bcceffbe-fb98-48de-a29d-51bd75e00c7a-kube-api-access-s8zv5\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:04 crc kubenswrapper[4972]: I0228 10:56:04.418755 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" event={"ID":"bcceffbe-fb98-48de-a29d-51bd75e00c7a","Type":"ContainerDied","Data":"76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9"} Feb 28 10:56:04 crc kubenswrapper[4972]: I0228 10:56:04.419092 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76f05b0b8a371ff663d87c397e8f16ce46a327564d7605bd215e532f327271d9" Feb 28 10:56:04 crc kubenswrapper[4972]: I0228 10:56:04.418835 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537936-lwr9x" Feb 28 10:56:04 crc kubenswrapper[4972]: I0228 10:56:04.902539 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537930-8wvgq"] Feb 28 10:56:04 crc kubenswrapper[4972]: I0228 10:56:04.908224 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537930-8wvgq"] Feb 28 10:56:05 crc kubenswrapper[4972]: I0228 10:56:05.436255 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerStarted","Data":"cb4af361fab42229b2e558c85328ba37dd3c80056980e527544a492724ed15de"} Feb 28 10:56:05 crc kubenswrapper[4972]: I0228 10:56:05.801226 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a163a1-a70c-4052-bc4c-d54cf60b4613" path="/var/lib/kubelet/pods/35a163a1-a70c-4052-bc4c-d54cf60b4613/volumes" Feb 28 10:56:06 crc kubenswrapper[4972]: I0228 10:56:06.445223 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:56:06 crc kubenswrapper[4972]: I0228 10:56:06.473394 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9256671819999998 podStartE2EDuration="7.473372607s" podCreationTimestamp="2026-02-28 10:55:59 +0000 UTC" firstStartedPulling="2026-02-28 10:56:00.237005052 +0000 UTC m=+1217.149012800" lastFinishedPulling="2026-02-28 10:56:04.784710487 +0000 UTC m=+1221.696718225" observedRunningTime="2026-02-28 10:56:06.467522304 +0000 UTC m=+1223.379530042" watchObservedRunningTime="2026-02-28 10:56:06.473372607 +0000 UTC m=+1223.385380345" Feb 28 10:56:06 crc kubenswrapper[4972]: I0228 10:56:06.984910 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:06 crc kubenswrapper[4972]: I0228 10:56:06.985525 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-log" containerID="cri-o://1c72d419d0c725671b6225463290a64838849df9bfd96d7ac9b36684ca917807" gracePeriod=30 Feb 28 10:56:06 crc kubenswrapper[4972]: I0228 10:56:06.985634 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-httpd" containerID="cri-o://ad5e9d261027f103027150480385caf8114daa4369d7c7c2646d66eaee8a2f94" gracePeriod=30 Feb 28 10:56:07 crc kubenswrapper[4972]: I0228 10:56:07.457650 4972 generic.go:334] "Generic (PLEG): container finished" podID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerID="1c72d419d0c725671b6225463290a64838849df9bfd96d7ac9b36684ca917807" exitCode=143 Feb 28 10:56:07 crc kubenswrapper[4972]: I0228 10:56:07.457743 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerDied","Data":"1c72d419d0c725671b6225463290a64838849df9bfd96d7ac9b36684ca917807"} Feb 28 10:56:07 crc kubenswrapper[4972]: I0228 10:56:07.918388 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:07 crc kubenswrapper[4972]: I0228 10:56:07.919252 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-log" containerID="cri-o://5e1a8e0c6c8ad4ffa9e50e451cfe899f0a2f0f316295b1f5df98c6b1089d1226" gracePeriod=30 Feb 28 10:56:07 crc kubenswrapper[4972]: I0228 10:56:07.919369 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-httpd" containerID="cri-o://ed4586be8bb58a2b22c320a60174e676c9f5a97c0ae37f6505a39f16bad92ba7" gracePeriod=30 Feb 28 10:56:08 crc kubenswrapper[4972]: I0228 10:56:08.467662 4972 generic.go:334] "Generic (PLEG): container finished" podID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerID="5e1a8e0c6c8ad4ffa9e50e451cfe899f0a2f0f316295b1f5df98c6b1089d1226" exitCode=143 Feb 28 10:56:08 crc kubenswrapper[4972]: I0228 10:56:08.467709 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerDied","Data":"5e1a8e0c6c8ad4ffa9e50e451cfe899f0a2f0f316295b1f5df98c6b1089d1226"} Feb 28 10:56:09 crc kubenswrapper[4972]: I0228 10:56:09.575679 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:09 crc kubenswrapper[4972]: I0228 10:56:09.575998 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-central-agent" containerID="cri-o://203b4deaec4bdfa1707b66a279e1b87d04e1077fa797daafd3535b78e039fa7b" gracePeriod=30 Feb 28 10:56:09 crc kubenswrapper[4972]: I0228 10:56:09.576101 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="proxy-httpd" containerID="cri-o://cb4af361fab42229b2e558c85328ba37dd3c80056980e527544a492724ed15de" gracePeriod=30 Feb 28 10:56:09 crc kubenswrapper[4972]: I0228 10:56:09.576132 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="sg-core" containerID="cri-o://8c66ad1568a2ef2d47315e2a1a2da762c4275d7a47c5f80598667d01eb85a82f" gracePeriod=30 Feb 28 10:56:09 crc kubenswrapper[4972]: I0228 10:56:09.576218 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-notification-agent" containerID="cri-o://2acd85bd4eb3d7b200e659fd3c24122828d037ed5fe7420bd3f2b4926ff7537c" gracePeriod=30 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.492687 4972 generic.go:334] "Generic (PLEG): container finished" podID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerID="ad5e9d261027f103027150480385caf8114daa4369d7c7c2646d66eaee8a2f94" exitCode=0 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.492790 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerDied","Data":"ad5e9d261027f103027150480385caf8114daa4369d7c7c2646d66eaee8a2f94"} Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.497729 4972 generic.go:334] "Generic (PLEG): container finished" podID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerID="cb4af361fab42229b2e558c85328ba37dd3c80056980e527544a492724ed15de" exitCode=0 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.497971 4972 generic.go:334] "Generic (PLEG): container finished" podID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerID="8c66ad1568a2ef2d47315e2a1a2da762c4275d7a47c5f80598667d01eb85a82f" exitCode=2 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.498074 4972 generic.go:334] "Generic (PLEG): container finished" podID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerID="2acd85bd4eb3d7b200e659fd3c24122828d037ed5fe7420bd3f2b4926ff7537c" exitCode=0 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.498187 4972 generic.go:334] "Generic (PLEG): container finished" podID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerID="203b4deaec4bdfa1707b66a279e1b87d04e1077fa797daafd3535b78e039fa7b" exitCode=0 Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.497796 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerDied","Data":"cb4af361fab42229b2e558c85328ba37dd3c80056980e527544a492724ed15de"} Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.498500 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerDied","Data":"8c66ad1568a2ef2d47315e2a1a2da762c4275d7a47c5f80598667d01eb85a82f"} Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.498604 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerDied","Data":"2acd85bd4eb3d7b200e659fd3c24122828d037ed5fe7420bd3f2b4926ff7537c"} Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.498691 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerDied","Data":"203b4deaec4bdfa1707b66a279e1b87d04e1077fa797daafd3535b78e039fa7b"} Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.570736 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.669756 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.705909 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706046 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706121 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706164 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b7wx\" (UniqueName: \"kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706279 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706368 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706400 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd\") pod \"467042a7-d82c-4b76-a3d9-b3db51b89471\" (UID: \"467042a7-d82c-4b76-a3d9-b3db51b89471\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.706525 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.707002 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.707413 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.716873 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts" (OuterVolumeSpecName: "scripts") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.716941 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx" (OuterVolumeSpecName: "kube-api-access-7b7wx") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "kube-api-access-7b7wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.787726 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.810391 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812696 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812752 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812829 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812891 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812921 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.812959 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.813063 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxtm4\" (UniqueName: \"kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4\") pod \"db0b990a-60c3-4d43-9f11-883155d37a5a\" (UID: \"db0b990a-60c3-4d43-9f11-883155d37a5a\") " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.814075 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.814097 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/467042a7-d82c-4b76-a3d9-b3db51b89471-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.814107 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.814118 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b7wx\" (UniqueName: \"kubernetes.io/projected/467042a7-d82c-4b76-a3d9-b3db51b89471-kube-api-access-7b7wx\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.818701 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.821988 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4" (OuterVolumeSpecName: "kube-api-access-vxtm4") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "kube-api-access-vxtm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.822391 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs" (OuterVolumeSpecName: "logs") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.869824 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts" (OuterVolumeSpecName: "scripts") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.869913 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.894804 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.909710 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917072 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917118 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917128 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917137 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db0b990a-60c3-4d43-9f11-883155d37a5a-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917160 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917169 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxtm4\" (UniqueName: \"kubernetes.io/projected/db0b990a-60c3-4d43-9f11-883155d37a5a-kube-api-access-vxtm4\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.917182 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.948935 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.957556 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:10 crc kubenswrapper[4972]: I0228 10:56:10.969498 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data" (OuterVolumeSpecName: "config-data") pod "db0b990a-60c3-4d43-9f11-883155d37a5a" (UID: "db0b990a-60c3-4d43-9f11-883155d37a5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.008584 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data" (OuterVolumeSpecName: "config-data") pod "467042a7-d82c-4b76-a3d9-b3db51b89471" (UID: "467042a7-d82c-4b76-a3d9-b3db51b89471"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.018904 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.018938 4972 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db0b990a-60c3-4d43-9f11-883155d37a5a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.018951 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.018964 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467042a7-d82c-4b76-a3d9-b3db51b89471-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.510758 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db0b990a-60c3-4d43-9f11-883155d37a5a","Type":"ContainerDied","Data":"bf02cc5f0126e963966b788c404a617d431712af189fad6061b19640b6c99db8"} Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.511265 4972 scope.go:117] "RemoveContainer" containerID="ad5e9d261027f103027150480385caf8114daa4369d7c7c2646d66eaee8a2f94" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.511341 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.521581 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"467042a7-d82c-4b76-a3d9-b3db51b89471","Type":"ContainerDied","Data":"418c286642128f3b7fa56e8049db8e35775f2f8fb5d0b4622e2bf1059a455c9f"} Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.521646 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.585172 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.597216 4972 scope.go:117] "RemoveContainer" containerID="1c72d419d0c725671b6225463290a64838849df9bfd96d7ac9b36684ca917807" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.602125 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.610582 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.619774 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.626620 4972 scope.go:117] "RemoveContainer" containerID="cb4af361fab42229b2e558c85328ba37dd3c80056980e527544a492724ed15de" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.654574 4972 scope.go:117] "RemoveContainer" containerID="8c66ad1568a2ef2d47315e2a1a2da762c4275d7a47c5f80598667d01eb85a82f" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.657664 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658165 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658227 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658310 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="proxy-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658376 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="proxy-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658433 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcceffbe-fb98-48de-a29d-51bd75e00c7a" containerName="oc" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658575 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcceffbe-fb98-48de-a29d-51bd75e00c7a" containerName="oc" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658630 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-log" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658679 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-log" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658747 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-notification-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658802 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-notification-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.658861 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="sg-core" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.658916 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="sg-core" Feb 28 10:56:11 crc kubenswrapper[4972]: E0228 10:56:11.659017 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-central-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659075 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-central-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659305 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="sg-core" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659416 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcceffbe-fb98-48de-a29d-51bd75e00c7a" containerName="oc" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659498 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659566 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" containerName="glance-log" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659629 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-notification-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659684 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="proxy-httpd" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.659742 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" containerName="ceilometer-central-agent" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.660783 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.666004 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.666068 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.671909 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.674365 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.683274 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.683574 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.692746 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.698947 4972 scope.go:117] "RemoveContainer" containerID="2acd85bd4eb3d7b200e659fd3c24122828d037ed5fe7420bd3f2b4926ff7537c" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.715119 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.794884 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.794947 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-config-data\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.794977 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795015 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795040 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795081 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795120 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9xm\" (UniqueName: \"kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795140 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s2kn\" (UniqueName: \"kubernetes.io/projected/e00dd1be-06c1-4c4a-979e-3cb562e7741e-kube-api-access-2s2kn\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795165 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795195 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795240 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795263 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-scripts\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795295 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795327 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-logs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.795361 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.816256 4972 scope.go:117] "RemoveContainer" containerID="203b4deaec4bdfa1707b66a279e1b87d04e1077fa797daafd3535b78e039fa7b" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.836192 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467042a7-d82c-4b76-a3d9-b3db51b89471" path="/var/lib/kubelet/pods/467042a7-d82c-4b76-a3d9-b3db51b89471/volumes" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.837122 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0b990a-60c3-4d43-9f11-883155d37a5a" path="/var/lib/kubelet/pods/db0b990a-60c3-4d43-9f11-883155d37a5a/volumes" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897011 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-logs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897068 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897143 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897170 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-config-data\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897189 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897230 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897254 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897269 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897296 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9xm\" (UniqueName: \"kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897311 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s2kn\" (UniqueName: \"kubernetes.io/projected/e00dd1be-06c1-4c4a-979e-3cb562e7741e-kube-api-access-2s2kn\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897354 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897380 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897431 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897449 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-scripts\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897517 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.897790 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.898223 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-logs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.903377 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.904517 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.904840 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.906920 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.908130 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e00dd1be-06c1-4c4a-979e-3cb562e7741e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.910354 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.910630 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.912862 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-config-data\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.912915 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-scripts\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.917428 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.923842 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00dd1be-06c1-4c4a-979e-3cb562e7741e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.928335 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s2kn\" (UniqueName: \"kubernetes.io/projected/e00dd1be-06c1-4c4a-979e-3cb562e7741e-kube-api-access-2s2kn\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.931961 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9xm\" (UniqueName: \"kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm\") pod \"ceilometer-0\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " pod="openstack/ceilometer-0" Feb 28 10:56:11 crc kubenswrapper[4972]: I0228 10:56:11.956911 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"e00dd1be-06c1-4c4a-979e-3cb562e7741e\") " pod="openstack/glance-default-external-api-0" Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.064290 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.086794 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.537225 4972 generic.go:334] "Generic (PLEG): container finished" podID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerID="ed4586be8bb58a2b22c320a60174e676c9f5a97c0ae37f6505a39f16bad92ba7" exitCode=0 Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.537508 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerDied","Data":"ed4586be8bb58a2b22c320a60174e676c9f5a97c0ae37f6505a39f16bad92ba7"} Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.665502 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 28 10:56:12 crc kubenswrapper[4972]: I0228 10:56:12.690387 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:13 crc kubenswrapper[4972]: I0228 10:56:13.594537 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerStarted","Data":"b701df631397fb7cfb3625068f688853d83fb446847d55ca36b5bab3b6fd75f5"} Feb 28 10:56:13 crc kubenswrapper[4972]: I0228 10:56:13.616751 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e00dd1be-06c1-4c4a-979e-3cb562e7741e","Type":"ContainerStarted","Data":"eb9ff027bb91fd055e3ca478ed0a8781945dd668ea60c5cb18b0cff7562dd4fe"} Feb 28 10:56:13 crc kubenswrapper[4972]: I0228 10:56:13.616807 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e00dd1be-06c1-4c4a-979e-3cb562e7741e","Type":"ContainerStarted","Data":"659d957f6b017eaca78812926b7d8132a6e73ccf5381d055e19e5957c894e19a"} Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.014801 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148207 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148257 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148321 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqvpv\" (UniqueName: \"kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148362 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148536 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148572 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148600 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148662 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs\") pod \"2ce78cca-b508-4c23-b331-21f1840ba47c\" (UID: \"2ce78cca-b508-4c23-b331-21f1840ba47c\") " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.148919 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs" (OuterVolumeSpecName: "logs") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.149127 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.152313 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.156284 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv" (OuterVolumeSpecName: "kube-api-access-vqvpv") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "kube-api-access-vqvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.158675 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.158830 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts" (OuterVolumeSpecName: "scripts") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.182868 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.212103 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data" (OuterVolumeSpecName: "config-data") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.213864 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ce78cca-b508-4c23-b331-21f1840ba47c" (UID: "2ce78cca-b508-4c23-b331-21f1840ba47c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251039 4972 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251076 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251089 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqvpv\" (UniqueName: \"kubernetes.io/projected/2ce78cca-b508-4c23-b331-21f1840ba47c-kube-api-access-vqvpv\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251138 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251150 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251159 4972 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2ce78cca-b508-4c23-b331-21f1840ba47c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.251168 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ce78cca-b508-4c23-b331-21f1840ba47c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.279570 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.354142 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.629351 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2ce78cca-b508-4c23-b331-21f1840ba47c","Type":"ContainerDied","Data":"e5000cce091bf3e88641440b6d8f6e4032745a604c00f297473e69cdbcf3dbcb"} Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.629417 4972 scope.go:117] "RemoveContainer" containerID="ed4586be8bb58a2b22c320a60174e676c9f5a97c0ae37f6505a39f16bad92ba7" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.629653 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.635343 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerStarted","Data":"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb"} Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.647796 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e00dd1be-06c1-4c4a-979e-3cb562e7741e","Type":"ContainerStarted","Data":"c80da299027cbf7ebd3c6da0f2920c796fb8cbc11f2cac4608124ad5000dee32"} Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.675688 4972 scope.go:117] "RemoveContainer" containerID="5e1a8e0c6c8ad4ffa9e50e451cfe899f0a2f0f316295b1f5df98c6b1089d1226" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.677622 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.711791 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.721922 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:14 crc kubenswrapper[4972]: E0228 10:56:14.722610 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-httpd" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.722634 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-httpd" Feb 28 10:56:14 crc kubenswrapper[4972]: E0228 10:56:14.722689 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-log" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.722697 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-log" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.722963 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-httpd" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.722992 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" containerName="glance-log" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.724644 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.727263 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.727535 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.727995 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.727964348 podStartE2EDuration="3.727964348s" podCreationTimestamp="2026-02-28 10:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:14.69621331 +0000 UTC m=+1231.608221048" watchObservedRunningTime="2026-02-28 10:56:14.727964348 +0000 UTC m=+1231.639972086" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.748592 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.867713 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.867782 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.867974 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.868156 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.868202 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.868521 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.868741 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-logs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.868829 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d44wz\" (UniqueName: \"kubernetes.io/projected/35925034-cd2f-4537-8225-0ab87fd3a47e-kube-api-access-d44wz\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.971743 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.971831 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.971993 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972074 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972150 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972244 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972338 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-logs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972251 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.972407 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d44wz\" (UniqueName: \"kubernetes.io/projected/35925034-cd2f-4537-8225-0ab87fd3a47e-kube-api-access-d44wz\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.973078 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-logs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.973210 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35925034-cd2f-4537-8225-0ab87fd3a47e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.977288 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.979738 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.981503 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:14 crc kubenswrapper[4972]: I0228 10:56:14.988367 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/35925034-cd2f-4537-8225-0ab87fd3a47e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.006752 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d44wz\" (UniqueName: \"kubernetes.io/projected/35925034-cd2f-4537-8225-0ab87fd3a47e-kube-api-access-d44wz\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.008314 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"35925034-cd2f-4537-8225-0ab87fd3a47e\") " pod="openstack/glance-default-internal-api-0" Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.049028 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:15 crc kubenswrapper[4972]: W0228 10:56:15.663160 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35925034_cd2f_4537_8225_0ab87fd3a47e.slice/crio-19107cef13715fb4dc5ca449fa7f007cb084219f95f53a76f120143e8c3aaba8 WatchSource:0}: Error finding container 19107cef13715fb4dc5ca449fa7f007cb084219f95f53a76f120143e8c3aaba8: Status 404 returned error can't find the container with id 19107cef13715fb4dc5ca449fa7f007cb084219f95f53a76f120143e8c3aaba8 Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.670223 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerStarted","Data":"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f"} Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.674787 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 28 10:56:15 crc kubenswrapper[4972]: I0228 10:56:15.802282 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ce78cca-b508-4c23-b331-21f1840ba47c" path="/var/lib/kubelet/pods/2ce78cca-b508-4c23-b331-21f1840ba47c/volumes" Feb 28 10:56:16 crc kubenswrapper[4972]: I0228 10:56:16.704926 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35925034-cd2f-4537-8225-0ab87fd3a47e","Type":"ContainerStarted","Data":"adb613a4e90472580c4aec2fb818adab60a38e9fc06603bd4ecfb15b4ee5f76d"} Feb 28 10:56:16 crc kubenswrapper[4972]: I0228 10:56:16.705309 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35925034-cd2f-4537-8225-0ab87fd3a47e","Type":"ContainerStarted","Data":"19107cef13715fb4dc5ca449fa7f007cb084219f95f53a76f120143e8c3aaba8"} Feb 28 10:56:17 crc kubenswrapper[4972]: I0228 10:56:17.720604 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerStarted","Data":"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49"} Feb 28 10:56:17 crc kubenswrapper[4972]: I0228 10:56:17.723746 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"35925034-cd2f-4537-8225-0ab87fd3a47e","Type":"ContainerStarted","Data":"ad8577ed0081b2d78612a83d8f9ff8019863a876289144b0f010baca24b2ece8"} Feb 28 10:56:17 crc kubenswrapper[4972]: I0228 10:56:17.758429 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.758411399 podStartE2EDuration="3.758411399s" podCreationTimestamp="2026-02-28 10:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:17.752980468 +0000 UTC m=+1234.664988216" watchObservedRunningTime="2026-02-28 10:56:17.758411399 +0000 UTC m=+1234.670419127" Feb 28 10:56:19 crc kubenswrapper[4972]: I0228 10:56:19.757378 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerStarted","Data":"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49"} Feb 28 10:56:19 crc kubenswrapper[4972]: I0228 10:56:19.758254 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:56:19 crc kubenswrapper[4972]: I0228 10:56:19.788117 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.699070492 podStartE2EDuration="8.788080543s" podCreationTimestamp="2026-02-28 10:56:11 +0000 UTC" firstStartedPulling="2026-02-28 10:56:12.720073451 +0000 UTC m=+1229.632081189" lastFinishedPulling="2026-02-28 10:56:18.809083502 +0000 UTC m=+1235.721091240" observedRunningTime="2026-02-28 10:56:19.784146473 +0000 UTC m=+1236.696154211" watchObservedRunningTime="2026-02-28 10:56:19.788080543 +0000 UTC m=+1236.700088281" Feb 28 10:56:20 crc kubenswrapper[4972]: I0228 10:56:20.769954 4972 generic.go:334] "Generic (PLEG): container finished" podID="6e7caa58-92af-435f-9d46-6f63ed19a6e6" containerID="a6cdca71b0d7d252124e9d485c53110feb4235f37dbb35cf332a1d2a62d6750b" exitCode=0 Feb 28 10:56:20 crc kubenswrapper[4972]: I0228 10:56:20.770035 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-56fjs" event={"ID":"6e7caa58-92af-435f-9d46-6f63ed19a6e6","Type":"ContainerDied","Data":"a6cdca71b0d7d252124e9d485c53110feb4235f37dbb35cf332a1d2a62d6750b"} Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.065392 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.067170 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.103984 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.114958 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.167263 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.256986 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2xm8\" (UniqueName: \"kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8\") pod \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.257113 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle\") pod \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.257211 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts\") pod \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.257372 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data\") pod \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\" (UID: \"6e7caa58-92af-435f-9d46-6f63ed19a6e6\") " Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.262763 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8" (OuterVolumeSpecName: "kube-api-access-b2xm8") pod "6e7caa58-92af-435f-9d46-6f63ed19a6e6" (UID: "6e7caa58-92af-435f-9d46-6f63ed19a6e6"). InnerVolumeSpecName "kube-api-access-b2xm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.262985 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts" (OuterVolumeSpecName: "scripts") pod "6e7caa58-92af-435f-9d46-6f63ed19a6e6" (UID: "6e7caa58-92af-435f-9d46-6f63ed19a6e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.284045 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data" (OuterVolumeSpecName: "config-data") pod "6e7caa58-92af-435f-9d46-6f63ed19a6e6" (UID: "6e7caa58-92af-435f-9d46-6f63ed19a6e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.287971 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e7caa58-92af-435f-9d46-6f63ed19a6e6" (UID: "6e7caa58-92af-435f-9d46-6f63ed19a6e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.359969 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.360008 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.360020 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e7caa58-92af-435f-9d46-6f63ed19a6e6-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.360031 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2xm8\" (UniqueName: \"kubernetes.io/projected/6e7caa58-92af-435f-9d46-6f63ed19a6e6-kube-api-access-b2xm8\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.794645 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-56fjs" event={"ID":"6e7caa58-92af-435f-9d46-6f63ed19a6e6","Type":"ContainerDied","Data":"04fd827ecb6a823ba4729f69fafc691bbbde95e40daeea43d53765382569b6b2"} Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.794699 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04fd827ecb6a823ba4729f69fafc691bbbde95e40daeea43d53765382569b6b2" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.794708 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-56fjs" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.795026 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.795066 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.931552 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 28 10:56:22 crc kubenswrapper[4972]: E0228 10:56:22.931941 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7caa58-92af-435f-9d46-6f63ed19a6e6" containerName="nova-cell0-conductor-db-sync" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.931959 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7caa58-92af-435f-9d46-6f63ed19a6e6" containerName="nova-cell0-conductor-db-sync" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.932133 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7caa58-92af-435f-9d46-6f63ed19a6e6" containerName="nova-cell0-conductor-db-sync" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.932702 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.936612 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-25pgq" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.936825 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.953941 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.983090 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.983161 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:22 crc kubenswrapper[4972]: I0228 10:56:22.983226 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4gst\" (UniqueName: \"kubernetes.io/projected/3e68715e-2416-4459-8c91-6368dbedb67a-kube-api-access-p4gst\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.085637 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.085687 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.085723 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4gst\" (UniqueName: \"kubernetes.io/projected/3e68715e-2416-4459-8c91-6368dbedb67a-kube-api-access-p4gst\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.089392 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.090986 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e68715e-2416-4459-8c91-6368dbedb67a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.102816 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4gst\" (UniqueName: \"kubernetes.io/projected/3e68715e-2416-4459-8c91-6368dbedb67a-kube-api-access-p4gst\") pod \"nova-cell0-conductor-0\" (UID: \"3e68715e-2416-4459-8c91-6368dbedb67a\") " pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.295577 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:23 crc kubenswrapper[4972]: W0228 10:56:23.795376 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e68715e_2416_4459_8c91_6368dbedb67a.slice/crio-5201b62435db11825c98020ae4d49cdd396afa6574d0628c8ed35e5b7e0fb6e6 WatchSource:0}: Error finding container 5201b62435db11825c98020ae4d49cdd396afa6574d0628c8ed35e5b7e0fb6e6: Status 404 returned error can't find the container with id 5201b62435db11825c98020ae4d49cdd396afa6574d0628c8ed35e5b7e0fb6e6 Feb 28 10:56:23 crc kubenswrapper[4972]: I0228 10:56:23.805693 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.805861 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.807304 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.819044 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3e68715e-2416-4459-8c91-6368dbedb67a","Type":"ContainerStarted","Data":"97d2205972747d3417342097419dc7faa2670f8532a3fe5784d2db7d23bc931b"} Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.819112 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3e68715e-2416-4459-8c91-6368dbedb67a","Type":"ContainerStarted","Data":"5201b62435db11825c98020ae4d49cdd396afa6574d0628c8ed35e5b7e0fb6e6"} Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.819245 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:24 crc kubenswrapper[4972]: I0228 10:56:24.856157 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.85613348 podStartE2EDuration="2.85613348s" podCreationTimestamp="2026-02-28 10:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:24.851427428 +0000 UTC m=+1241.763435166" watchObservedRunningTime="2026-02-28 10:56:24.85613348 +0000 UTC m=+1241.768141228" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.049828 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.050238 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.085154 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.103860 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.830105 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:25 crc kubenswrapper[4972]: I0228 10:56:25.830159 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:27 crc kubenswrapper[4972]: I0228 10:56:27.806687 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:27 crc kubenswrapper[4972]: I0228 10:56:27.844819 4972 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 28 10:56:27 crc kubenswrapper[4972]: I0228 10:56:27.880694 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.330716 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.863978 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lrmpd"] Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.865232 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.867750 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.869138 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 28 10:56:28 crc kubenswrapper[4972]: I0228 10:56:28.881330 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrmpd"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.034243 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.034442 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.034508 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf7fk\" (UniqueName: \"kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.034831 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.036689 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.037176 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.039094 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.065840 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.122694 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.124219 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.126310 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.138821 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.138871 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.138912 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sczcx\" (UniqueName: \"kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.139013 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.139062 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.139120 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.139140 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.139166 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf7fk\" (UniqueName: \"kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.145366 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.153188 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.167800 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.177093 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.217554 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf7fk\" (UniqueName: \"kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk\") pod \"nova-cell0-cell-mapping-lrmpd\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248003 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sczcx\" (UniqueName: \"kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248065 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248110 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248143 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrvb2\" (UniqueName: \"kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248203 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248282 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248351 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.248382 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.251962 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.256682 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.266174 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.288253 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sczcx\" (UniqueName: \"kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx\") pod \"nova-api-0\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.353141 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.353871 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrvb2\" (UniqueName: \"kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.354070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.354128 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.354162 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.354836 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.360182 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.388159 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.388874 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.389061 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.390447 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.390485 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.390563 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.391198 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.403725 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.435226 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrvb2\" (UniqueName: \"kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2\") pod \"nova-metadata-0\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.474558 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.475909 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.487834 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.502853 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.664852 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.666835 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p7zm\" (UniqueName: \"kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.666993 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667120 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667202 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667230 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667341 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667398 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk8vp\" (UniqueName: \"kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667446 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.667493 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.679328 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773199 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773251 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84t5g\" (UniqueName: \"kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773299 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p7zm\" (UniqueName: \"kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773339 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773374 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773404 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773419 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773447 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773487 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773507 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk8vp\" (UniqueName: \"kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773529 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.773547 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.774360 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.781341 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.783749 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.786125 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.786372 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.790202 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p7zm\" (UniqueName: \"kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm\") pod \"nova-scheduler-0\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.792201 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.792489 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.798373 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk8vp\" (UniqueName: \"kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp\") pod \"dnsmasq-dns-757b4f8459-jwt5h\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.876859 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.878025 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.878070 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84t5g\" (UniqueName: \"kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.885775 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.887851 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.904108 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84t5g\" (UniqueName: \"kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g\") pod \"nova-cell1-novncproxy-0\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:29 crc kubenswrapper[4972]: I0228 10:56:29.908782 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.025117 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.066793 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.188451 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.330151 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrmpd"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.447425 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:30 crc kubenswrapper[4972]: W0228 10:56:30.460614 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0d367da_ce08_405c_8a2d_2d7bbe02e5d9.slice/crio-fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef WatchSource:0}: Error finding container fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef: Status 404 returned error can't find the container with id fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.529277 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-522pb"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.532162 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.535788 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.537281 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.543934 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-522pb"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.579111 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.699681 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpsbc\" (UniqueName: \"kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.699998 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.700146 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.700297 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: W0228 10:56:30.702840 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a47f25b_cb53_4ed8_8cec_78deb919494e.slice/crio-fc26b6411ac963137f7af2e50ad766d42d2bd206c1f424bda1665137f9280a53 WatchSource:0}: Error finding container fc26b6411ac963137f7af2e50ad766d42d2bd206c1f424bda1665137f9280a53: Status 404 returned error can't find the container with id fc26b6411ac963137f7af2e50ad766d42d2bd206c1f424bda1665137f9280a53 Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.705293 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.713985 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:56:30 crc kubenswrapper[4972]: W0228 10:56:30.720634 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8212c91a_9a0e_4873_b7ca_2547d5aa515a.slice/crio-e3eb544d61f0feee6f07810eeab22304833e86f40745b44950af3c86a1e4abaa WatchSource:0}: Error finding container e3eb544d61f0feee6f07810eeab22304833e86f40745b44950af3c86a1e4abaa: Status 404 returned error can't find the container with id e3eb544d61f0feee6f07810eeab22304833e86f40745b44950af3c86a1e4abaa Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.802154 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpsbc\" (UniqueName: \"kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.802293 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.802339 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.802385 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.806375 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.807622 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.807889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.822035 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpsbc\" (UniqueName: \"kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc\") pod \"nova-cell1-conductor-db-sync-522pb\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.863276 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.882132 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerStarted","Data":"a2e02ff9b03815afec638097e16298f7fea5b92c612cc1283617eaa7d96d717e"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.883267 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerStarted","Data":"fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.885930 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bae35ca5-1401-4d44-b2ca-71df17a88318","Type":"ContainerStarted","Data":"1211c00a75afbc9f67c9eac08e0ebf06259fad81f23a0ddf6af8837a4cb7144d"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.887937 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" event={"ID":"8212c91a-9a0e-4873-b7ca-2547d5aa515a","Type":"ContainerStarted","Data":"e3eb544d61f0feee6f07810eeab22304833e86f40745b44950af3c86a1e4abaa"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.888941 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a47f25b-cb53-4ed8-8cec-78deb919494e","Type":"ContainerStarted","Data":"fc26b6411ac963137f7af2e50ad766d42d2bd206c1f424bda1665137f9280a53"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.890013 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrmpd" event={"ID":"67229e04-2f99-4419-b160-a8871c61c886","Type":"ContainerStarted","Data":"c54d631349c50bf9a494cdd8ad00bcc19510126580db34ee1f024e06388ed8fc"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.890040 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrmpd" event={"ID":"67229e04-2f99-4419-b160-a8871c61c886","Type":"ContainerStarted","Data":"41665e8f373b0b73ec7e7b15fde11df6259ee630d34a765a4eec61503cf1c82d"} Feb 28 10:56:30 crc kubenswrapper[4972]: I0228 10:56:30.917019 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lrmpd" podStartSLOduration=2.916993774 podStartE2EDuration="2.916993774s" podCreationTimestamp="2026-02-28 10:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:30.907493768 +0000 UTC m=+1247.819501506" watchObservedRunningTime="2026-02-28 10:56:30.916993774 +0000 UTC m=+1247.829001532" Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.385159 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-522pb"] Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.912583 4972 generic.go:334] "Generic (PLEG): container finished" podID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerID="28f5ab92e31d20609202fb169287845d7d75861f5eac59b57889b9984222140c" exitCode=0 Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.912698 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" event={"ID":"8212c91a-9a0e-4873-b7ca-2547d5aa515a","Type":"ContainerDied","Data":"28f5ab92e31d20609202fb169287845d7d75861f5eac59b57889b9984222140c"} Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.917809 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-522pb" event={"ID":"39eca2d3-a87e-47f8-9d3b-ee07a5969701","Type":"ContainerStarted","Data":"7a0084846917d0aaf5389f62c60cd27ee77ce90eb5d14d810ba3f5add64801b7"} Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.917893 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-522pb" event={"ID":"39eca2d3-a87e-47f8-9d3b-ee07a5969701","Type":"ContainerStarted","Data":"fa4d28bdff9e498cbc2558d63e856e3c8624f00154b9885675b47038e122a147"} Feb 28 10:56:31 crc kubenswrapper[4972]: I0228 10:56:31.967703 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-522pb" podStartSLOduration=1.967683277 podStartE2EDuration="1.967683277s" podCreationTimestamp="2026-02-28 10:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:31.962890613 +0000 UTC m=+1248.874898351" watchObservedRunningTime="2026-02-28 10:56:31.967683277 +0000 UTC m=+1248.879691015" Feb 28 10:56:32 crc kubenswrapper[4972]: I0228 10:56:32.745021 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:32 crc kubenswrapper[4972]: I0228 10:56:32.758137 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.969772 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a47f25b-cb53-4ed8-8cec-78deb919494e","Type":"ContainerStarted","Data":"aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.976330 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerStarted","Data":"02e38bb812c755a3ec7dd3777a912d0c8a52a9917b86e3bb2537d97529610e60"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.976372 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerStarted","Data":"c2e9f049968c8fba5f7682f173c606b5117969ad6139b15a7fdb95718e1937bc"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.981424 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerStarted","Data":"567e79466eff3c51dff9964630c86e829e206b558c26a76c56c7d7d4336a075c"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.981549 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerStarted","Data":"3d06d7a543157f8e61090743f82cc4472b8d43ab796062e2f9f7d939df6f13be"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.981700 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-log" containerID="cri-o://3d06d7a543157f8e61090743f82cc4472b8d43ab796062e2f9f7d939df6f13be" gracePeriod=30 Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.981786 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-metadata" containerID="cri-o://567e79466eff3c51dff9964630c86e829e206b558c26a76c56c7d7d4336a075c" gracePeriod=30 Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.986766 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bae35ca5-1401-4d44-b2ca-71df17a88318","Type":"ContainerStarted","Data":"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683"} Feb 28 10:56:35 crc kubenswrapper[4972]: I0228 10:56:35.986890 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bae35ca5-1401-4d44-b2ca-71df17a88318" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683" gracePeriod=30 Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.002076 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" event={"ID":"8212c91a-9a0e-4873-b7ca-2547d5aa515a","Type":"ContainerStarted","Data":"5f5288931eab8244b943d57472b2fecc0ec7c4254595258eba9a03a7acb4da8e"} Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.002353 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.007010 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.961899865 podStartE2EDuration="7.006987414s" podCreationTimestamp="2026-02-28 10:56:29 +0000 UTC" firstStartedPulling="2026-02-28 10:56:30.709599262 +0000 UTC m=+1247.621607000" lastFinishedPulling="2026-02-28 10:56:34.754686811 +0000 UTC m=+1251.666694549" observedRunningTime="2026-02-28 10:56:35.989359901 +0000 UTC m=+1252.901367639" watchObservedRunningTime="2026-02-28 10:56:36.006987414 +0000 UTC m=+1252.918995152" Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.020569 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7148203950000003 podStartE2EDuration="7.020543642s" podCreationTimestamp="2026-02-28 10:56:29 +0000 UTC" firstStartedPulling="2026-02-28 10:56:30.464641851 +0000 UTC m=+1247.376649589" lastFinishedPulling="2026-02-28 10:56:34.770365088 +0000 UTC m=+1251.682372836" observedRunningTime="2026-02-28 10:56:36.019605806 +0000 UTC m=+1252.931613554" watchObservedRunningTime="2026-02-28 10:56:36.020543642 +0000 UTC m=+1252.932551420" Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.048835 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.492821806 podStartE2EDuration="7.048817352s" podCreationTimestamp="2026-02-28 10:56:29 +0000 UTC" firstStartedPulling="2026-02-28 10:56:30.214638749 +0000 UTC m=+1247.126646487" lastFinishedPulling="2026-02-28 10:56:34.770634295 +0000 UTC m=+1251.682642033" observedRunningTime="2026-02-28 10:56:36.042176136 +0000 UTC m=+1252.954183904" watchObservedRunningTime="2026-02-28 10:56:36.048817352 +0000 UTC m=+1252.960825090" Feb 28 10:56:36 crc kubenswrapper[4972]: I0228 10:56:36.069100 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.897492206 podStartE2EDuration="7.069080348s" podCreationTimestamp="2026-02-28 10:56:29 +0000 UTC" firstStartedPulling="2026-02-28 10:56:30.58314365 +0000 UTC m=+1247.495151388" lastFinishedPulling="2026-02-28 10:56:34.754731792 +0000 UTC m=+1251.666739530" observedRunningTime="2026-02-28 10:56:36.059833369 +0000 UTC m=+1252.971841117" watchObservedRunningTime="2026-02-28 10:56:36.069080348 +0000 UTC m=+1252.981088086" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.013610 4972 generic.go:334] "Generic (PLEG): container finished" podID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerID="567e79466eff3c51dff9964630c86e829e206b558c26a76c56c7d7d4336a075c" exitCode=0 Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.014119 4972 generic.go:334] "Generic (PLEG): container finished" podID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerID="3d06d7a543157f8e61090743f82cc4472b8d43ab796062e2f9f7d939df6f13be" exitCode=143 Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.013696 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerDied","Data":"567e79466eff3c51dff9964630c86e829e206b558c26a76c56c7d7d4336a075c"} Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.014366 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerDied","Data":"3d06d7a543157f8e61090743f82cc4472b8d43ab796062e2f9f7d939df6f13be"} Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.014415 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9","Type":"ContainerDied","Data":"fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef"} Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.014440 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa46195fb26b08b083324dc68359c02274c2efe22115933dd2b7b22e912d2fef" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.076115 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.111272 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" podStartSLOduration=8.111249483 podStartE2EDuration="8.111249483s" podCreationTimestamp="2026-02-28 10:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:36.078926222 +0000 UTC m=+1252.990933980" watchObservedRunningTime="2026-02-28 10:56:37.111249483 +0000 UTC m=+1254.023257221" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.187788 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs\") pod \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.187908 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") pod \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.188039 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrvb2\" (UniqueName: \"kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2\") pod \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.188133 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle\") pod \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.188385 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs" (OuterVolumeSpecName: "logs") pod "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" (UID: "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.188939 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.209094 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2" (OuterVolumeSpecName: "kube-api-access-lrvb2") pod "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" (UID: "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9"). InnerVolumeSpecName "kube-api-access-lrvb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:37 crc kubenswrapper[4972]: E0228 10:56:37.233365 4972 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data podName:a0d367da-ce08-405c-8a2d-2d7bbe02e5d9 nodeName:}" failed. No retries permitted until 2026-02-28 10:56:37.733330692 +0000 UTC m=+1254.645338450 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data") pod "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" (UID: "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9") : error deleting /var/lib/kubelet/pods/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9/volume-subpaths: remove /var/lib/kubelet/pods/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9/volume-subpaths: no such file or directory Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.237567 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" (UID: "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.291831 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.292038 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrvb2\" (UniqueName: \"kubernetes.io/projected/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-kube-api-access-lrvb2\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.801661 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") pod \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\" (UID: \"a0d367da-ce08-405c-8a2d-2d7bbe02e5d9\") " Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.806410 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data" (OuterVolumeSpecName: "config-data") pod "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" (UID: "a0d367da-ce08-405c-8a2d-2d7bbe02e5d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:37 crc kubenswrapper[4972]: I0228 10:56:37.905269 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.028764 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.084341 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.094792 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.112781 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:38 crc kubenswrapper[4972]: E0228 10:56:38.113438 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-log" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.113549 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-log" Feb 28 10:56:38 crc kubenswrapper[4972]: E0228 10:56:38.113580 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-metadata" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.113586 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-metadata" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.113805 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-metadata" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.113828 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" containerName="nova-metadata-log" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.114843 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.118224 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.118409 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.135618 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.313700 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.314114 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kb78\" (UniqueName: \"kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.314326 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.314435 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.314481 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.416029 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.416117 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.416145 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.416203 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.416225 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kb78\" (UniqueName: \"kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.417213 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.421007 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.421213 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.422089 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.450727 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kb78\" (UniqueName: \"kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78\") pod \"nova-metadata-0\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " pod="openstack/nova-metadata-0" Feb 28 10:56:38 crc kubenswrapper[4972]: I0228 10:56:38.734487 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:39 crc kubenswrapper[4972]: W0228 10:56:39.275275 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ded963_b61f_4a51_8a80_86d9ca007de1.slice/crio-933d60936b94aea0cea734f04ed859c07b600a2cb4149119fd410d8e4689bbba WatchSource:0}: Error finding container 933d60936b94aea0cea734f04ed859c07b600a2cb4149119fd410d8e4689bbba: Status 404 returned error can't find the container with id 933d60936b94aea0cea734f04ed859c07b600a2cb4149119fd410d8e4689bbba Feb 28 10:56:39 crc kubenswrapper[4972]: I0228 10:56:39.277652 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:39 crc kubenswrapper[4972]: I0228 10:56:39.353643 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:56:39 crc kubenswrapper[4972]: I0228 10:56:39.353733 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:56:39 crc kubenswrapper[4972]: I0228 10:56:39.809569 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d367da-ce08-405c-8a2d-2d7bbe02e5d9" path="/var/lib/kubelet/pods/a0d367da-ce08-405c-8a2d-2d7bbe02e5d9/volumes" Feb 28 10:56:39 crc kubenswrapper[4972]: I0228 10:56:39.910554 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.027583 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.062008 4972 generic.go:334] "Generic (PLEG): container finished" podID="67229e04-2f99-4419-b160-a8871c61c886" containerID="c54d631349c50bf9a494cdd8ad00bcc19510126580db34ee1f024e06388ed8fc" exitCode=0 Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.062095 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrmpd" event={"ID":"67229e04-2f99-4419-b160-a8871c61c886","Type":"ContainerDied","Data":"c54d631349c50bf9a494cdd8ad00bcc19510126580db34ee1f024e06388ed8fc"} Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.067741 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.067845 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.075275 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerStarted","Data":"cde51fc559cb997bbcf4c780e303f5e1f48403a946ac46b315e8db9ed46cc838"} Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.075324 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerStarted","Data":"dccf4685b94a7da49092a39b997089e19a49e89af944f9a4e901c32605351408"} Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.075338 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerStarted","Data":"933d60936b94aea0cea734f04ed859c07b600a2cb4149119fd410d8e4689bbba"} Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.109245 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.137087 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.137058216 podStartE2EDuration="2.137058216s" podCreationTimestamp="2026-02-28 10:56:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:40.117602842 +0000 UTC m=+1257.029610590" watchObservedRunningTime="2026-02-28 10:56:40.137058216 +0000 UTC m=+1257.049065954" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.139558 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.139885 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="dnsmasq-dns" containerID="cri-o://bf541cc76cf02c659b3a50beeae0508bd468413e17e045507719169ba8bc0733" gracePeriod=10 Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.437400 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:56:40 crc kubenswrapper[4972]: I0228 10:56:40.437509 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.088028 4972 generic.go:334] "Generic (PLEG): container finished" podID="9499375a-dad8-43ee-af2f-02963356cf50" containerID="bf541cc76cf02c659b3a50beeae0508bd468413e17e045507719169ba8bc0733" exitCode=0 Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.088095 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" event={"ID":"9499375a-dad8-43ee-af2f-02963356cf50","Type":"ContainerDied","Data":"bf541cc76cf02c659b3a50beeae0508bd468413e17e045507719169ba8bc0733"} Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.088672 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" event={"ID":"9499375a-dad8-43ee-af2f-02963356cf50","Type":"ContainerDied","Data":"24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487"} Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.088720 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24e753541548a878142496f18905b837235ca4aa32f61820c8f98c2c7c3f0487" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.123102 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.151759 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.292872 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.293380 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.293450 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.293533 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t858g\" (UniqueName: \"kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.293568 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.293621 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config\") pod \"9499375a-dad8-43ee-af2f-02963356cf50\" (UID: \"9499375a-dad8-43ee-af2f-02963356cf50\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.301718 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g" (OuterVolumeSpecName: "kube-api-access-t858g") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "kube-api-access-t858g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.361798 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.378733 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config" (OuterVolumeSpecName: "config") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.379398 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.379655 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.383306 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9499375a-dad8-43ee-af2f-02963356cf50" (UID: "9499375a-dad8-43ee-af2f-02963356cf50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396588 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396621 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396630 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396638 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396648 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9499375a-dad8-43ee-af2f-02963356cf50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.396655 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t858g\" (UniqueName: \"kubernetes.io/projected/9499375a-dad8-43ee-af2f-02963356cf50-kube-api-access-t858g\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.538235 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.599126 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts\") pod \"67229e04-2f99-4419-b160-a8871c61c886\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.599221 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data\") pod \"67229e04-2f99-4419-b160-a8871c61c886\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.599269 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf7fk\" (UniqueName: \"kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk\") pod \"67229e04-2f99-4419-b160-a8871c61c886\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.599477 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle\") pod \"67229e04-2f99-4419-b160-a8871c61c886\" (UID: \"67229e04-2f99-4419-b160-a8871c61c886\") " Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.604247 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk" (OuterVolumeSpecName: "kube-api-access-zf7fk") pod "67229e04-2f99-4419-b160-a8871c61c886" (UID: "67229e04-2f99-4419-b160-a8871c61c886"). InnerVolumeSpecName "kube-api-access-zf7fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.604953 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts" (OuterVolumeSpecName: "scripts") pod "67229e04-2f99-4419-b160-a8871c61c886" (UID: "67229e04-2f99-4419-b160-a8871c61c886"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.635942 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data" (OuterVolumeSpecName: "config-data") pod "67229e04-2f99-4419-b160-a8871c61c886" (UID: "67229e04-2f99-4419-b160-a8871c61c886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.637318 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67229e04-2f99-4419-b160-a8871c61c886" (UID: "67229e04-2f99-4419-b160-a8871c61c886"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.701728 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.701771 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.701788 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf7fk\" (UniqueName: \"kubernetes.io/projected/67229e04-2f99-4419-b160-a8871c61c886-kube-api-access-zf7fk\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:41 crc kubenswrapper[4972]: I0228 10:56:41.701803 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67229e04-2f99-4419-b160-a8871c61c886-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.100376 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lrmpd" event={"ID":"67229e04-2f99-4419-b160-a8871c61c886","Type":"ContainerDied","Data":"41665e8f373b0b73ec7e7b15fde11df6259ee630d34a765a4eec61503cf1c82d"} Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.100434 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41665e8f373b0b73ec7e7b15fde11df6259ee630d34a765a4eec61503cf1c82d" Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.100400 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-n4z6j" Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.101530 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lrmpd" Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.121711 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.136098 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.146198 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-n4z6j"] Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.313557 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.313911 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-log" containerID="cri-o://c2e9f049968c8fba5f7682f173c606b5117969ad6139b15a7fdb95718e1937bc" gracePeriod=30 Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.314448 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-api" containerID="cri-o://02e38bb812c755a3ec7dd3777a912d0c8a52a9917b86e3bb2537d97529610e60" gracePeriod=30 Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.328285 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.352761 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.353006 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-log" containerID="cri-o://dccf4685b94a7da49092a39b997089e19a49e89af944f9a4e901c32605351408" gracePeriod=30 Feb 28 10:56:42 crc kubenswrapper[4972]: I0228 10:56:42.353143 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-metadata" containerID="cri-o://cde51fc559cb997bbcf4c780e303f5e1f48403a946ac46b315e8db9ed46cc838" gracePeriod=30 Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.119978 4972 generic.go:334] "Generic (PLEG): container finished" podID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerID="c2e9f049968c8fba5f7682f173c606b5117969ad6139b15a7fdb95718e1937bc" exitCode=143 Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.120079 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerDied","Data":"c2e9f049968c8fba5f7682f173c606b5117969ad6139b15a7fdb95718e1937bc"} Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.135013 4972 generic.go:334] "Generic (PLEG): container finished" podID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerID="cde51fc559cb997bbcf4c780e303f5e1f48403a946ac46b315e8db9ed46cc838" exitCode=0 Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.135046 4972 generic.go:334] "Generic (PLEG): container finished" podID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerID="dccf4685b94a7da49092a39b997089e19a49e89af944f9a4e901c32605351408" exitCode=143 Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.135220 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerDied","Data":"cde51fc559cb997bbcf4c780e303f5e1f48403a946ac46b315e8db9ed46cc838"} Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.135307 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerDied","Data":"dccf4685b94a7da49092a39b997089e19a49e89af944f9a4e901c32605351408"} Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.293371 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.339998 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle\") pod \"a0ded963-b61f-4a51-8a80-86d9ca007de1\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.340063 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs\") pod \"a0ded963-b61f-4a51-8a80-86d9ca007de1\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.340223 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kb78\" (UniqueName: \"kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78\") pod \"a0ded963-b61f-4a51-8a80-86d9ca007de1\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.340284 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs\") pod \"a0ded963-b61f-4a51-8a80-86d9ca007de1\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.340382 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data\") pod \"a0ded963-b61f-4a51-8a80-86d9ca007de1\" (UID: \"a0ded963-b61f-4a51-8a80-86d9ca007de1\") " Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.350933 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs" (OuterVolumeSpecName: "logs") pod "a0ded963-b61f-4a51-8a80-86d9ca007de1" (UID: "a0ded963-b61f-4a51-8a80-86d9ca007de1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.357647 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78" (OuterVolumeSpecName: "kube-api-access-2kb78") pod "a0ded963-b61f-4a51-8a80-86d9ca007de1" (UID: "a0ded963-b61f-4a51-8a80-86d9ca007de1"). InnerVolumeSpecName "kube-api-access-2kb78". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.384618 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ded963-b61f-4a51-8a80-86d9ca007de1" (UID: "a0ded963-b61f-4a51-8a80-86d9ca007de1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.384668 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data" (OuterVolumeSpecName: "config-data") pod "a0ded963-b61f-4a51-8a80-86d9ca007de1" (UID: "a0ded963-b61f-4a51-8a80-86d9ca007de1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.413370 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a0ded963-b61f-4a51-8a80-86d9ca007de1" (UID: "a0ded963-b61f-4a51-8a80-86d9ca007de1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.442923 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.442959 4972 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.442975 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kb78\" (UniqueName: \"kubernetes.io/projected/a0ded963-b61f-4a51-8a80-86d9ca007de1-kube-api-access-2kb78\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.442983 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ded963-b61f-4a51-8a80-86d9ca007de1-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.442993 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ded963-b61f-4a51-8a80-86d9ca007de1-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:43 crc kubenswrapper[4972]: I0228 10:56:43.804176 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9499375a-dad8-43ee-af2f-02963356cf50" path="/var/lib/kubelet/pods/9499375a-dad8-43ee-af2f-02963356cf50/volumes" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.148111 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerName="nova-scheduler-scheduler" containerID="cri-o://aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" gracePeriod=30 Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.148439 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.149299 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ded963-b61f-4a51-8a80-86d9ca007de1","Type":"ContainerDied","Data":"933d60936b94aea0cea734f04ed859c07b600a2cb4149119fd410d8e4689bbba"} Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.149338 4972 scope.go:117] "RemoveContainer" containerID="cde51fc559cb997bbcf4c780e303f5e1f48403a946ac46b315e8db9ed46cc838" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.178226 4972 scope.go:117] "RemoveContainer" containerID="dccf4685b94a7da49092a39b997089e19a49e89af944f9a4e901c32605351408" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.189539 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.207846 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227353 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:44 crc kubenswrapper[4972]: E0228 10:56:44.227835 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-metadata" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227856 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-metadata" Feb 28 10:56:44 crc kubenswrapper[4972]: E0228 10:56:44.227875 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67229e04-2f99-4419-b160-a8871c61c886" containerName="nova-manage" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227882 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="67229e04-2f99-4419-b160-a8871c61c886" containerName="nova-manage" Feb 28 10:56:44 crc kubenswrapper[4972]: E0228 10:56:44.227895 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-log" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227901 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-log" Feb 28 10:56:44 crc kubenswrapper[4972]: E0228 10:56:44.227930 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="init" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227937 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="init" Feb 28 10:56:44 crc kubenswrapper[4972]: E0228 10:56:44.227952 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="dnsmasq-dns" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.227959 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="dnsmasq-dns" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.228157 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-log" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.228175 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="67229e04-2f99-4419-b160-a8871c61c886" containerName="nova-manage" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.228184 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9499375a-dad8-43ee-af2f-02963356cf50" containerName="dnsmasq-dns" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.228199 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" containerName="nova-metadata-metadata" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.229295 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.232745 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.232803 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.239202 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.259484 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.259537 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w29d\" (UniqueName: \"kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.259589 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.259673 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.259691 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.362222 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.362297 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.362402 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.362438 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w29d\" (UniqueName: \"kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.362526 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.364827 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.370833 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.370872 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.371773 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.381989 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w29d\" (UniqueName: \"kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d\") pod \"nova-metadata-0\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " pod="openstack/nova-metadata-0" Feb 28 10:56:44 crc kubenswrapper[4972]: I0228 10:56:44.547512 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:56:45 crc kubenswrapper[4972]: E0228 10:56:45.071001 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 28 10:56:45 crc kubenswrapper[4972]: E0228 10:56:45.078082 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 28 10:56:45 crc kubenswrapper[4972]: E0228 10:56:45.080091 4972 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 28 10:56:45 crc kubenswrapper[4972]: E0228 10:56:45.080183 4972 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerName="nova-scheduler-scheduler" Feb 28 10:56:45 crc kubenswrapper[4972]: I0228 10:56:45.086970 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:56:45 crc kubenswrapper[4972]: I0228 10:56:45.166323 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerStarted","Data":"f66115d1a169dff011b58bccba8987b74bb84f49d3a51204bd68d4d70b845fd1"} Feb 28 10:56:45 crc kubenswrapper[4972]: I0228 10:56:45.806162 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ded963-b61f-4a51-8a80-86d9ca007de1" path="/var/lib/kubelet/pods/a0ded963-b61f-4a51-8a80-86d9ca007de1/volumes" Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.175257 4972 generic.go:334] "Generic (PLEG): container finished" podID="39eca2d3-a87e-47f8-9d3b-ee07a5969701" containerID="7a0084846917d0aaf5389f62c60cd27ee77ce90eb5d14d810ba3f5add64801b7" exitCode=0 Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.175320 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-522pb" event={"ID":"39eca2d3-a87e-47f8-9d3b-ee07a5969701","Type":"ContainerDied","Data":"7a0084846917d0aaf5389f62c60cd27ee77ce90eb5d14d810ba3f5add64801b7"} Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.177558 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerStarted","Data":"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024"} Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.177609 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerStarted","Data":"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a"} Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.244020 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.244000286 podStartE2EDuration="2.244000286s" podCreationTimestamp="2026-02-28 10:56:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:46.235768996 +0000 UTC m=+1263.147776734" watchObservedRunningTime="2026-02-28 10:56:46.244000286 +0000 UTC m=+1263.156008024" Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.580818 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:46 crc kubenswrapper[4972]: I0228 10:56:46.581689 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" containerName="kube-state-metrics" containerID="cri-o://2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272" gracePeriod=30 Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.208088 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.208386 4972 generic.go:334] "Generic (PLEG): container finished" podID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" containerID="2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272" exitCode=2 Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.208416 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"daacfde0-2575-4448-a11f-0eb9a1b2a1cb","Type":"ContainerDied","Data":"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272"} Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.208939 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"daacfde0-2575-4448-a11f-0eb9a1b2a1cb","Type":"ContainerDied","Data":"7a609c22878b138327aaf45b4b6c4de8a7002dd66a529eff4831fca5d528f4bd"} Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.208962 4972 scope.go:117] "RemoveContainer" containerID="2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.212316 4972 generic.go:334] "Generic (PLEG): container finished" podID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerID="aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" exitCode=0 Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.212412 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a47f25b-cb53-4ed8-8cec-78deb919494e","Type":"ContainerDied","Data":"aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce"} Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.218745 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scdw9\" (UniqueName: \"kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9\") pod \"daacfde0-2575-4448-a11f-0eb9a1b2a1cb\" (UID: \"daacfde0-2575-4448-a11f-0eb9a1b2a1cb\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.223097 4972 generic.go:334] "Generic (PLEG): container finished" podID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerID="02e38bb812c755a3ec7dd3777a912d0c8a52a9917b86e3bb2537d97529610e60" exitCode=0 Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.223340 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerDied","Data":"02e38bb812c755a3ec7dd3777a912d0c8a52a9917b86e3bb2537d97529610e60"} Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.240753 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9" (OuterVolumeSpecName: "kube-api-access-scdw9") pod "daacfde0-2575-4448-a11f-0eb9a1b2a1cb" (UID: "daacfde0-2575-4448-a11f-0eb9a1b2a1cb"). InnerVolumeSpecName "kube-api-access-scdw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.276394 4972 scope.go:117] "RemoveContainer" containerID="2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272" Feb 28 10:56:47 crc kubenswrapper[4972]: E0228 10:56:47.278105 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272\": container with ID starting with 2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272 not found: ID does not exist" containerID="2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.278213 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272"} err="failed to get container status \"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272\": rpc error: code = NotFound desc = could not find container \"2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272\": container with ID starting with 2ab229deb1ef63ac78ab13601a2689f1f950018747653f8175001a7d90017272 not found: ID does not exist" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.321676 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scdw9\" (UniqueName: \"kubernetes.io/projected/daacfde0-2575-4448-a11f-0eb9a1b2a1cb-kube-api-access-scdw9\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.331538 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.375699 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.525938 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data\") pod \"4a47f25b-cb53-4ed8-8cec-78deb919494e\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526285 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sczcx\" (UniqueName: \"kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx\") pod \"dbedd435-7f95-4aa4-bcc3-18682c150975\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526357 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle\") pod \"dbedd435-7f95-4aa4-bcc3-18682c150975\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526407 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p7zm\" (UniqueName: \"kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm\") pod \"4a47f25b-cb53-4ed8-8cec-78deb919494e\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526431 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle\") pod \"4a47f25b-cb53-4ed8-8cec-78deb919494e\" (UID: \"4a47f25b-cb53-4ed8-8cec-78deb919494e\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526514 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs\") pod \"dbedd435-7f95-4aa4-bcc3-18682c150975\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.526548 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data\") pod \"dbedd435-7f95-4aa4-bcc3-18682c150975\" (UID: \"dbedd435-7f95-4aa4-bcc3-18682c150975\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.528020 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs" (OuterVolumeSpecName: "logs") pod "dbedd435-7f95-4aa4-bcc3-18682c150975" (UID: "dbedd435-7f95-4aa4-bcc3-18682c150975"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.536151 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm" (OuterVolumeSpecName: "kube-api-access-7p7zm") pod "4a47f25b-cb53-4ed8-8cec-78deb919494e" (UID: "4a47f25b-cb53-4ed8-8cec-78deb919494e"). InnerVolumeSpecName "kube-api-access-7p7zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.555995 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx" (OuterVolumeSpecName: "kube-api-access-sczcx") pod "dbedd435-7f95-4aa4-bcc3-18682c150975" (UID: "dbedd435-7f95-4aa4-bcc3-18682c150975"). InnerVolumeSpecName "kube-api-access-sczcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.560920 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data" (OuterVolumeSpecName: "config-data") pod "dbedd435-7f95-4aa4-bcc3-18682c150975" (UID: "dbedd435-7f95-4aa4-bcc3-18682c150975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.563647 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a47f25b-cb53-4ed8-8cec-78deb919494e" (UID: "4a47f25b-cb53-4ed8-8cec-78deb919494e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.568773 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbedd435-7f95-4aa4-bcc3-18682c150975" (UID: "dbedd435-7f95-4aa4-bcc3-18682c150975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.586087 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data" (OuterVolumeSpecName: "config-data") pod "4a47f25b-cb53-4ed8-8cec-78deb919494e" (UID: "4a47f25b-cb53-4ed8-8cec-78deb919494e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.603971 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.628679 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sczcx\" (UniqueName: \"kubernetes.io/projected/dbedd435-7f95-4aa4-bcc3-18682c150975-kube-api-access-sczcx\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.628960 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.629021 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p7zm\" (UniqueName: \"kubernetes.io/projected/4a47f25b-cb53-4ed8-8cec-78deb919494e-kube-api-access-7p7zm\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.629098 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.629154 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbedd435-7f95-4aa4-bcc3-18682c150975-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.629206 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbedd435-7f95-4aa4-bcc3-18682c150975-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.629264 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a47f25b-cb53-4ed8-8cec-78deb919494e-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.730186 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle\") pod \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.730324 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts\") pod \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.730348 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data\") pod \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.730539 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpsbc\" (UniqueName: \"kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc\") pod \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\" (UID: \"39eca2d3-a87e-47f8-9d3b-ee07a5969701\") " Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.738347 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc" (OuterVolumeSpecName: "kube-api-access-zpsbc") pod "39eca2d3-a87e-47f8-9d3b-ee07a5969701" (UID: "39eca2d3-a87e-47f8-9d3b-ee07a5969701"). InnerVolumeSpecName "kube-api-access-zpsbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.738781 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts" (OuterVolumeSpecName: "scripts") pod "39eca2d3-a87e-47f8-9d3b-ee07a5969701" (UID: "39eca2d3-a87e-47f8-9d3b-ee07a5969701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.768149 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39eca2d3-a87e-47f8-9d3b-ee07a5969701" (UID: "39eca2d3-a87e-47f8-9d3b-ee07a5969701"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.782135 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data" (OuterVolumeSpecName: "config-data") pod "39eca2d3-a87e-47f8-9d3b-ee07a5969701" (UID: "39eca2d3-a87e-47f8-9d3b-ee07a5969701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.833071 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpsbc\" (UniqueName: \"kubernetes.io/projected/39eca2d3-a87e-47f8-9d3b-ee07a5969701-kube-api-access-zpsbc\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.833314 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.833408 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:47 crc kubenswrapper[4972]: I0228 10:56:47.833508 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39eca2d3-a87e-47f8-9d3b-ee07a5969701-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.233838 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-522pb" event={"ID":"39eca2d3-a87e-47f8-9d3b-ee07a5969701","Type":"ContainerDied","Data":"fa4d28bdff9e498cbc2558d63e856e3c8624f00154b9885675b47038e122a147"} Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.233862 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-522pb" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.233880 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa4d28bdff9e498cbc2558d63e856e3c8624f00154b9885675b47038e122a147" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.235088 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.236759 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a47f25b-cb53-4ed8-8cec-78deb919494e","Type":"ContainerDied","Data":"fc26b6411ac963137f7af2e50ad766d42d2bd206c1f424bda1665137f9280a53"} Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.236843 4972 scope.go:117] "RemoveContainer" containerID="aa009211dc29b8b94f443b8b6b16779d481116ce3f976a2354eb367ea2d731ce" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.237101 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.242853 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dbedd435-7f95-4aa4-bcc3-18682c150975","Type":"ContainerDied","Data":"a2e02ff9b03815afec638097e16298f7fea5b92c612cc1283617eaa7d96d717e"} Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.242902 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.271331 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.280612 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.284661 4972 scope.go:117] "RemoveContainer" containerID="02e38bb812c755a3ec7dd3777a912d0c8a52a9917b86e3bb2537d97529610e60" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.301875 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.321485 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.337730 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: E0228 10:56:48.338292 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-log" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338317 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-log" Feb 28 10:56:48 crc kubenswrapper[4972]: E0228 10:56:48.338337 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-api" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338345 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-api" Feb 28 10:56:48 crc kubenswrapper[4972]: E0228 10:56:48.338369 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39eca2d3-a87e-47f8-9d3b-ee07a5969701" containerName="nova-cell1-conductor-db-sync" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338378 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="39eca2d3-a87e-47f8-9d3b-ee07a5969701" containerName="nova-cell1-conductor-db-sync" Feb 28 10:56:48 crc kubenswrapper[4972]: E0228 10:56:48.338394 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" containerName="kube-state-metrics" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338405 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" containerName="kube-state-metrics" Feb 28 10:56:48 crc kubenswrapper[4972]: E0228 10:56:48.338425 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerName="nova-scheduler-scheduler" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338432 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerName="nova-scheduler-scheduler" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338685 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="39eca2d3-a87e-47f8-9d3b-ee07a5969701" containerName="nova-cell1-conductor-db-sync" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338758 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-log" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338776 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" containerName="kube-state-metrics" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338785 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" containerName="nova-scheduler-scheduler" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.338797 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" containerName="nova-api-api" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.339572 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.342303 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.344851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl6rj\" (UniqueName: \"kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.344970 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.345030 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.347008 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.348301 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.354394 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.376223 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.377611 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.383890 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.384005 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.388946 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.400510 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.423220 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.439291 4972 scope.go:117] "RemoveContainer" containerID="c2e9f049968c8fba5f7682f173c606b5117969ad6139b15a7fdb95718e1937bc" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.439418 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.450963 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.451097 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl6rj\" (UniqueName: \"kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.451161 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.455974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.456745 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.458472 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.472332 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.481494 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl6rj\" (UniqueName: \"kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj\") pod \"nova-scheduler-0\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.485472 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.488871 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.515779 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559625 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559666 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559701 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559724 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559746 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49n8d\" (UniqueName: \"kubernetes.io/projected/15239450-dea0-4aab-8fe7-1d891d57afab-kube-api-access-49n8d\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559764 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559836 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdkqm\" (UniqueName: \"kubernetes.io/projected/3563bc2b-2c86-4824-b238-e8a60f45c60e-kube-api-access-wdkqm\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559853 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5kks\" (UniqueName: \"kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.559905 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.560173 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.560374 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662601 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662677 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662721 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662762 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662815 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49n8d\" (UniqueName: \"kubernetes.io/projected/15239450-dea0-4aab-8fe7-1d891d57afab-kube-api-access-49n8d\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662842 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662933 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdkqm\" (UniqueName: \"kubernetes.io/projected/3563bc2b-2c86-4824-b238-e8a60f45c60e-kube-api-access-wdkqm\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662965 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5kks\" (UniqueName: \"kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.662997 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.663032 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.663084 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.663282 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.667674 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.668508 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.668698 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.669183 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.671166 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3563bc2b-2c86-4824-b238-e8a60f45c60e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.678494 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.679787 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/15239450-dea0-4aab-8fe7-1d891d57afab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.680156 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49n8d\" (UniqueName: \"kubernetes.io/projected/15239450-dea0-4aab-8fe7-1d891d57afab-kube-api-access-49n8d\") pod \"kube-state-metrics-0\" (UID: \"15239450-dea0-4aab-8fe7-1d891d57afab\") " pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.681970 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdkqm\" (UniqueName: \"kubernetes.io/projected/3563bc2b-2c86-4824-b238-e8a60f45c60e-kube-api-access-wdkqm\") pod \"nova-cell1-conductor-0\" (UID: \"3563bc2b-2c86-4824-b238-e8a60f45c60e\") " pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.686266 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5kks\" (UniqueName: \"kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks\") pod \"nova-api-0\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " pod="openstack/nova-api-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.687833 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.688889 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.721812 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.838296 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.839314 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-central-agent" containerID="cri-o://85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb" gracePeriod=30 Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.840092 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="proxy-httpd" containerID="cri-o://75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49" gracePeriod=30 Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.840375 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-notification-agent" containerID="cri-o://5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f" gracePeriod=30 Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.840427 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="sg-core" containerID="cri-o://61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49" gracePeriod=30 Feb 28 10:56:48 crc kubenswrapper[4972]: I0228 10:56:48.874792 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.240844 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.258301 4972 generic.go:334] "Generic (PLEG): container finished" podID="8a642b98-aefe-4490-b483-1b8f798a327c" containerID="75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49" exitCode=0 Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.258335 4972 generic.go:334] "Generic (PLEG): container finished" podID="8a642b98-aefe-4490-b483-1b8f798a327c" containerID="61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49" exitCode=2 Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.258375 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerDied","Data":"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49"} Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.258399 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerDied","Data":"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49"} Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.259397 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18465c59-c525-48fd-8898-3ae76bd54d11","Type":"ContainerStarted","Data":"7d14d47232d853f70f13b97f0d2b626a4ca22d04bbcfb69ae01460a8b3be0204"} Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.352193 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.382665 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 28 10:56:49 crc kubenswrapper[4972]: W0228 10:56:49.385251 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3563bc2b_2c86_4824_b238_e8a60f45c60e.slice/crio-0a0abb99997916c8ea03f8f60ecc42747b681eb0e0045c8807c95f5a2a02b57a WatchSource:0}: Error finding container 0a0abb99997916c8ea03f8f60ecc42747b681eb0e0045c8807c95f5a2a02b57a: Status 404 returned error can't find the container with id 0a0abb99997916c8ea03f8f60ecc42747b681eb0e0045c8807c95f5a2a02b57a Feb 28 10:56:49 crc kubenswrapper[4972]: W0228 10:56:49.519098 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aabf8aa_17d0_4e2a_ae12_d667e5545ca2.slice/crio-42caf911a7c67242494c06cd9ece6a39ff0991e4a7b5bb5ccdb8b251800cae29 WatchSource:0}: Error finding container 42caf911a7c67242494c06cd9ece6a39ff0991e4a7b5bb5ccdb8b251800cae29: Status 404 returned error can't find the container with id 42caf911a7c67242494c06cd9ece6a39ff0991e4a7b5bb5ccdb8b251800cae29 Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.526213 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.550163 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.551248 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.802271 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a47f25b-cb53-4ed8-8cec-78deb919494e" path="/var/lib/kubelet/pods/4a47f25b-cb53-4ed8-8cec-78deb919494e/volumes" Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.802769 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daacfde0-2575-4448-a11f-0eb9a1b2a1cb" path="/var/lib/kubelet/pods/daacfde0-2575-4448-a11f-0eb9a1b2a1cb/volumes" Feb 28 10:56:49 crc kubenswrapper[4972]: I0228 10:56:49.803285 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbedd435-7f95-4aa4-bcc3-18682c150975" path="/var/lib/kubelet/pods/dbedd435-7f95-4aa4-bcc3-18682c150975/volumes" Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.279589 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"15239450-dea0-4aab-8fe7-1d891d57afab","Type":"ContainerStarted","Data":"4584a66fbf9e44615d7ad2d68984976c30aad67dee2ec0d09e9eb3bd3d59a43e"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.284834 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerStarted","Data":"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.284869 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerStarted","Data":"42caf911a7c67242494c06cd9ece6a39ff0991e4a7b5bb5ccdb8b251800cae29"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.301940 4972 generic.go:334] "Generic (PLEG): container finished" podID="8a642b98-aefe-4490-b483-1b8f798a327c" containerID="85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb" exitCode=0 Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.302018 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerDied","Data":"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.317035 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18465c59-c525-48fd-8898-3ae76bd54d11","Type":"ContainerStarted","Data":"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.339157 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3563bc2b-2c86-4824-b238-e8a60f45c60e","Type":"ContainerStarted","Data":"996b91b155285cba6e11dba76c69e27ea2ebc103eaff2d121b30163600d3da2c"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.339217 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3563bc2b-2c86-4824-b238-e8a60f45c60e","Type":"ContainerStarted","Data":"0a0abb99997916c8ea03f8f60ecc42747b681eb0e0045c8807c95f5a2a02b57a"} Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.339271 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.365118 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.365097978 podStartE2EDuration="2.365097978s" podCreationTimestamp="2026-02-28 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:50.356405905 +0000 UTC m=+1267.268413643" watchObservedRunningTime="2026-02-28 10:56:50.365097978 +0000 UTC m=+1267.277105706" Feb 28 10:56:50 crc kubenswrapper[4972]: I0228 10:56:50.396481 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.396437183 podStartE2EDuration="2.396437183s" podCreationTimestamp="2026-02-28 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:50.384411257 +0000 UTC m=+1267.296418995" watchObservedRunningTime="2026-02-28 10:56:50.396437183 +0000 UTC m=+1267.308444921" Feb 28 10:56:51 crc kubenswrapper[4972]: I0228 10:56:51.349634 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"15239450-dea0-4aab-8fe7-1d891d57afab","Type":"ContainerStarted","Data":"e52bbd46e598c981b3be3f42d376b6d9708f530c724c09fe1b30550fa078fdcd"} Feb 28 10:56:51 crc kubenswrapper[4972]: I0228 10:56:51.350030 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 28 10:56:51 crc kubenswrapper[4972]: I0228 10:56:51.352496 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerStarted","Data":"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec"} Feb 28 10:56:51 crc kubenswrapper[4972]: I0228 10:56:51.379217 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.904849181 podStartE2EDuration="3.379192709s" podCreationTimestamp="2026-02-28 10:56:48 +0000 UTC" firstStartedPulling="2026-02-28 10:56:49.366241872 +0000 UTC m=+1266.278249610" lastFinishedPulling="2026-02-28 10:56:49.8405854 +0000 UTC m=+1266.752593138" observedRunningTime="2026-02-28 10:56:51.369384005 +0000 UTC m=+1268.281391743" watchObservedRunningTime="2026-02-28 10:56:51.379192709 +0000 UTC m=+1268.291200457" Feb 28 10:56:51 crc kubenswrapper[4972]: I0228 10:56:51.405741 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.405714819 podStartE2EDuration="3.405714819s" podCreationTimestamp="2026-02-28 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:56:51.394942819 +0000 UTC m=+1268.306950557" watchObservedRunningTime="2026-02-28 10:56:51.405714819 +0000 UTC m=+1268.317722557" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.124341 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.151741 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.151794 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.151849 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.151890 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr9xm\" (UniqueName: \"kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.152020 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.152046 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.152074 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd\") pod \"8a642b98-aefe-4490-b483-1b8f798a327c\" (UID: \"8a642b98-aefe-4490-b483-1b8f798a327c\") " Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.152932 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.158759 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.160366 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm" (OuterVolumeSpecName: "kube-api-access-hr9xm") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "kube-api-access-hr9xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.166109 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts" (OuterVolumeSpecName: "scripts") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.228422 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.250977 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254433 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254493 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254506 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254518 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254529 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a642b98-aefe-4490-b483-1b8f798a327c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.254539 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr9xm\" (UniqueName: \"kubernetes.io/projected/8a642b98-aefe-4490-b483-1b8f798a327c-kube-api-access-hr9xm\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.275971 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data" (OuterVolumeSpecName: "config-data") pod "8a642b98-aefe-4490-b483-1b8f798a327c" (UID: "8a642b98-aefe-4490-b483-1b8f798a327c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.355687 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a642b98-aefe-4490-b483-1b8f798a327c-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.364183 4972 generic.go:334] "Generic (PLEG): container finished" podID="8a642b98-aefe-4490-b483-1b8f798a327c" containerID="5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f" exitCode=0 Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.364272 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerDied","Data":"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f"} Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.364333 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a642b98-aefe-4490-b483-1b8f798a327c","Type":"ContainerDied","Data":"b701df631397fb7cfb3625068f688853d83fb446847d55ca36b5bab3b6fd75f5"} Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.364358 4972 scope.go:117] "RemoveContainer" containerID="75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.364541 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.384770 4972 scope.go:117] "RemoveContainer" containerID="61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.409475 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.420983 4972 scope.go:117] "RemoveContainer" containerID="5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.442420 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.470815 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.471614 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-central-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.472487 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-central-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.472605 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="proxy-httpd" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.472683 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="proxy-httpd" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.472777 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-notification-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.472849 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-notification-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.472943 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="sg-core" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.473020 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="sg-core" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.473346 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-notification-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.477017 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="ceilometer-central-agent" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.477239 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="proxy-httpd" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.477497 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" containerName="sg-core" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.475429 4972 scope.go:117] "RemoveContainer" containerID="85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.480251 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.480955 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.483479 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.483660 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.483704 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.524223 4972 scope.go:117] "RemoveContainer" containerID="75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.524811 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49\": container with ID starting with 75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49 not found: ID does not exist" containerID="75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.524898 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49"} err="failed to get container status \"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49\": rpc error: code = NotFound desc = could not find container \"75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49\": container with ID starting with 75355c58708d8a5376309db3c9f5f8adf23bc370b585b5cd368e570bc31bbc49 not found: ID does not exist" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.524973 4972 scope.go:117] "RemoveContainer" containerID="61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.525532 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49\": container with ID starting with 61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49 not found: ID does not exist" containerID="61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.525584 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49"} err="failed to get container status \"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49\": rpc error: code = NotFound desc = could not find container \"61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49\": container with ID starting with 61429ec22e7a8ff6e1d9404d09790a9058a5fa6e42470bacfd9a4ba0c977ac49 not found: ID does not exist" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.525619 4972 scope.go:117] "RemoveContainer" containerID="5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.525928 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f\": container with ID starting with 5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f not found: ID does not exist" containerID="5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.526044 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f"} err="failed to get container status \"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f\": rpc error: code = NotFound desc = could not find container \"5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f\": container with ID starting with 5b62ba50049ca35aad649850942a8e174cd198423f45ee9d4c251e9b6645e90f not found: ID does not exist" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.526142 4972 scope.go:117] "RemoveContainer" containerID="85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb" Feb 28 10:56:52 crc kubenswrapper[4972]: E0228 10:56:52.526489 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb\": container with ID starting with 85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb not found: ID does not exist" containerID="85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.526585 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb"} err="failed to get container status \"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb\": rpc error: code = NotFound desc = could not find container \"85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb\": container with ID starting with 85f3f4bd8da5eac9f7b72f1ba3210a0bbe364482055069fb7feb7c012c2085bb not found: ID does not exist" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561127 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561190 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561266 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561288 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561322 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561355 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptphw\" (UniqueName: \"kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561382 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.561409 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.663129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.663541 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.663694 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.663918 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.664554 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.665246 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.665369 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptphw\" (UniqueName: \"kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.665441 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.665506 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.665899 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.668989 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.670161 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.670287 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.670806 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.679984 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.683764 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptphw\" (UniqueName: \"kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw\") pod \"ceilometer-0\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " pod="openstack/ceilometer-0" Feb 28 10:56:52 crc kubenswrapper[4972]: I0228 10:56:52.811065 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:56:53 crc kubenswrapper[4972]: I0228 10:56:53.381384 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:56:53 crc kubenswrapper[4972]: W0228 10:56:53.398689 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99923c8e_fce0_44fb_b2b1_ef432ceefe7d.slice/crio-a477d5e7f5dbec9666a8828b4d6bd2d8eade0404b16e204082109d3c1b48a87a WatchSource:0}: Error finding container a477d5e7f5dbec9666a8828b4d6bd2d8eade0404b16e204082109d3c1b48a87a: Status 404 returned error can't find the container with id a477d5e7f5dbec9666a8828b4d6bd2d8eade0404b16e204082109d3c1b48a87a Feb 28 10:56:53 crc kubenswrapper[4972]: I0228 10:56:53.688793 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 28 10:56:53 crc kubenswrapper[4972]: I0228 10:56:53.805254 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a642b98-aefe-4490-b483-1b8f798a327c" path="/var/lib/kubelet/pods/8a642b98-aefe-4490-b483-1b8f798a327c/volumes" Feb 28 10:56:54 crc kubenswrapper[4972]: I0228 10:56:54.394034 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerStarted","Data":"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9"} Feb 28 10:56:54 crc kubenswrapper[4972]: I0228 10:56:54.394378 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerStarted","Data":"a477d5e7f5dbec9666a8828b4d6bd2d8eade0404b16e204082109d3c1b48a87a"} Feb 28 10:56:54 crc kubenswrapper[4972]: I0228 10:56:54.548007 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 28 10:56:54 crc kubenswrapper[4972]: I0228 10:56:54.548287 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 28 10:56:55 crc kubenswrapper[4972]: I0228 10:56:55.431531 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerStarted","Data":"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8"} Feb 28 10:56:55 crc kubenswrapper[4972]: I0228 10:56:55.563658 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:56:55 crc kubenswrapper[4972]: I0228 10:56:55.564010 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:56:56 crc kubenswrapper[4972]: I0228 10:56:56.449632 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerStarted","Data":"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e"} Feb 28 10:56:57 crc kubenswrapper[4972]: I0228 10:56:57.461238 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerStarted","Data":"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66"} Feb 28 10:56:57 crc kubenswrapper[4972]: I0228 10:56:57.461686 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:56:57 crc kubenswrapper[4972]: I0228 10:56:57.489354 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.702333108 podStartE2EDuration="5.489336459s" podCreationTimestamp="2026-02-28 10:56:52 +0000 UTC" firstStartedPulling="2026-02-28 10:56:53.404622774 +0000 UTC m=+1270.316630512" lastFinishedPulling="2026-02-28 10:56:57.191626125 +0000 UTC m=+1274.103633863" observedRunningTime="2026-02-28 10:56:57.4818559 +0000 UTC m=+1274.393863668" watchObservedRunningTime="2026-02-28 10:56:57.489336459 +0000 UTC m=+1274.401344197" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.688176 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.727349 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.742052 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.776222 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.875930 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:56:58 crc kubenswrapper[4972]: I0228 10:56:58.876009 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:56:59 crc kubenswrapper[4972]: I0228 10:56:59.284081 4972 scope.go:117] "RemoveContainer" containerID="bbc503054e17d6b71e191c0971b7edc76ab645da74b4e9634eef321953ea4704" Feb 28 10:56:59 crc kubenswrapper[4972]: I0228 10:56:59.540434 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 28 10:56:59 crc kubenswrapper[4972]: I0228 10:56:59.959737 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:56:59 crc kubenswrapper[4972]: I0228 10:56:59.959776 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:04 crc kubenswrapper[4972]: I0228 10:57:04.561904 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 28 10:57:04 crc kubenswrapper[4972]: I0228 10:57:04.571974 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 28 10:57:04 crc kubenswrapper[4972]: I0228 10:57:04.582900 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 28 10:57:05 crc kubenswrapper[4972]: I0228 10:57:05.545385 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.504115 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.548552 4972 generic.go:334] "Generic (PLEG): container finished" podID="bae35ca5-1401-4d44-b2ca-71df17a88318" containerID="d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683" exitCode=137 Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.548639 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.548684 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bae35ca5-1401-4d44-b2ca-71df17a88318","Type":"ContainerDied","Data":"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683"} Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.548721 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bae35ca5-1401-4d44-b2ca-71df17a88318","Type":"ContainerDied","Data":"1211c00a75afbc9f67c9eac08e0ebf06259fad81f23a0ddf6af8837a4cb7144d"} Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.548743 4972 scope.go:117] "RemoveContainer" containerID="d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.587849 4972 scope.go:117] "RemoveContainer" containerID="d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683" Feb 28 10:57:06 crc kubenswrapper[4972]: E0228 10:57:06.588356 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683\": container with ID starting with d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683 not found: ID does not exist" containerID="d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.588390 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683"} err="failed to get container status \"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683\": rpc error: code = NotFound desc = could not find container \"d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683\": container with ID starting with d02e7e865b80c92bb77dc4755db6b36751336e55c58768c5fb82b6dd7570d683 not found: ID does not exist" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.679953 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle\") pod \"bae35ca5-1401-4d44-b2ca-71df17a88318\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.680006 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data\") pod \"bae35ca5-1401-4d44-b2ca-71df17a88318\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.680325 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84t5g\" (UniqueName: \"kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g\") pod \"bae35ca5-1401-4d44-b2ca-71df17a88318\" (UID: \"bae35ca5-1401-4d44-b2ca-71df17a88318\") " Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.687828 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g" (OuterVolumeSpecName: "kube-api-access-84t5g") pod "bae35ca5-1401-4d44-b2ca-71df17a88318" (UID: "bae35ca5-1401-4d44-b2ca-71df17a88318"). InnerVolumeSpecName "kube-api-access-84t5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.715127 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae35ca5-1401-4d44-b2ca-71df17a88318" (UID: "bae35ca5-1401-4d44-b2ca-71df17a88318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.721585 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data" (OuterVolumeSpecName: "config-data") pod "bae35ca5-1401-4d44-b2ca-71df17a88318" (UID: "bae35ca5-1401-4d44-b2ca-71df17a88318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.782608 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84t5g\" (UniqueName: \"kubernetes.io/projected/bae35ca5-1401-4d44-b2ca-71df17a88318-kube-api-access-84t5g\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.782645 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.782657 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae35ca5-1401-4d44-b2ca-71df17a88318-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.892506 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.905090 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.923163 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:57:06 crc kubenswrapper[4972]: E0228 10:57:06.923990 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae35ca5-1401-4d44-b2ca-71df17a88318" containerName="nova-cell1-novncproxy-novncproxy" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.924030 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae35ca5-1401-4d44-b2ca-71df17a88318" containerName="nova-cell1-novncproxy-novncproxy" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.924553 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae35ca5-1401-4d44-b2ca-71df17a88318" containerName="nova-cell1-novncproxy-novncproxy" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.925739 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.928313 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.932577 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.938025 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:57:06 crc kubenswrapper[4972]: I0228 10:57:06.940564 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.089341 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.089481 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.089583 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.089701 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.089802 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmkdt\" (UniqueName: \"kubernetes.io/projected/70d1c7cf-2d6e-4076-b7cc-8752e5172008-kube-api-access-zmkdt\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.191284 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.191424 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmkdt\" (UniqueName: \"kubernetes.io/projected/70d1c7cf-2d6e-4076-b7cc-8752e5172008-kube-api-access-zmkdt\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.191575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.191640 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.191688 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.197756 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.204256 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.204650 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.206535 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/70d1c7cf-2d6e-4076-b7cc-8752e5172008-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.213422 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmkdt\" (UniqueName: \"kubernetes.io/projected/70d1c7cf-2d6e-4076-b7cc-8752e5172008-kube-api-access-zmkdt\") pod \"nova-cell1-novncproxy-0\" (UID: \"70d1c7cf-2d6e-4076-b7cc-8752e5172008\") " pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.245930 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:07 crc kubenswrapper[4972]: W0228 10:57:07.755170 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70d1c7cf_2d6e_4076_b7cc_8752e5172008.slice/crio-7793717f97be684bc348396a4259bc4fa98013493919a5ae24c8e5e5aae4903e WatchSource:0}: Error finding container 7793717f97be684bc348396a4259bc4fa98013493919a5ae24c8e5e5aae4903e: Status 404 returned error can't find the container with id 7793717f97be684bc348396a4259bc4fa98013493919a5ae24c8e5e5aae4903e Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.759579 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 28 10:57:07 crc kubenswrapper[4972]: I0228 10:57:07.810346 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae35ca5-1401-4d44-b2ca-71df17a88318" path="/var/lib/kubelet/pods/bae35ca5-1401-4d44-b2ca-71df17a88318/volumes" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.606511 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d1c7cf-2d6e-4076-b7cc-8752e5172008","Type":"ContainerStarted","Data":"aaea5adbb85cb5f77cb7009c90f1c8104c288e3f841e1c40e8966f3e97bc585c"} Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.606563 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"70d1c7cf-2d6e-4076-b7cc-8752e5172008","Type":"ContainerStarted","Data":"7793717f97be684bc348396a4259bc4fa98013493919a5ae24c8e5e5aae4903e"} Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.630817 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.630797951 podStartE2EDuration="2.630797951s" podCreationTimestamp="2026-02-28 10:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:08.627313014 +0000 UTC m=+1285.539320752" watchObservedRunningTime="2026-02-28 10:57:08.630797951 +0000 UTC m=+1285.542805689" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.880528 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.881011 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.881391 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.881432 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.884336 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 28 10:57:08 crc kubenswrapper[4972]: I0228 10:57:08.886527 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.124906 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.127084 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.171529 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.242964 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.243029 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.243059 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.243076 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs87g\" (UniqueName: \"kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.243101 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.243734 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346076 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346207 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346240 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346267 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346283 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs87g\" (UniqueName: \"kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.346314 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.347507 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.348032 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.348580 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.348696 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.349179 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.370528 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs87g\" (UniqueName: \"kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g\") pod \"dnsmasq-dns-89c5cd4d5-fk8xl\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:09 crc kubenswrapper[4972]: I0228 10:57:09.463288 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:10 crc kubenswrapper[4972]: I0228 10:57:10.014207 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:57:10 crc kubenswrapper[4972]: I0228 10:57:10.632366 4972 generic.go:334] "Generic (PLEG): container finished" podID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerID="d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267" exitCode=0 Feb 28 10:57:10 crc kubenswrapper[4972]: I0228 10:57:10.632503 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" event={"ID":"7fbf7692-27d2-4415-bd1d-ee7aace8f117","Type":"ContainerDied","Data":"d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267"} Feb 28 10:57:10 crc kubenswrapper[4972]: I0228 10:57:10.632570 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" event={"ID":"7fbf7692-27d2-4415-bd1d-ee7aace8f117","Type":"ContainerStarted","Data":"dec37811ed815d5910673588b49c5056de5f50f2e64e56ada95e81cca652e343"} Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.591291 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.592815 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-central-agent" containerID="cri-o://101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9" gracePeriod=30 Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.592891 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="sg-core" containerID="cri-o://2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e" gracePeriod=30 Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.592921 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-notification-agent" containerID="cri-o://7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8" gracePeriod=30 Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.592989 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="proxy-httpd" containerID="cri-o://b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66" gracePeriod=30 Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.608516 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.205:3000/\": EOF" Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.643490 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" event={"ID":"7fbf7692-27d2-4415-bd1d-ee7aace8f117","Type":"ContainerStarted","Data":"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401"} Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.644353 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.668552 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" podStartSLOduration=2.668526557 podStartE2EDuration="2.668526557s" podCreationTimestamp="2026-02-28 10:57:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:11.665511882 +0000 UTC m=+1288.577519630" watchObservedRunningTime="2026-02-28 10:57:11.668526557 +0000 UTC m=+1288.580534295" Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.979774 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.982998 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-log" containerID="cri-o://a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2" gracePeriod=30 Feb 28 10:57:11 crc kubenswrapper[4972]: I0228 10:57:11.983187 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-api" containerID="cri-o://78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec" gracePeriod=30 Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.246029 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.655871 4972 generic.go:334] "Generic (PLEG): container finished" podID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerID="b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66" exitCode=0 Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.655909 4972 generic.go:334] "Generic (PLEG): container finished" podID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerID="2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e" exitCode=2 Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.655922 4972 generic.go:334] "Generic (PLEG): container finished" podID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerID="101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9" exitCode=0 Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.655997 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerDied","Data":"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66"} Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.656025 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerDied","Data":"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e"} Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.656036 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerDied","Data":"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9"} Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.657744 4972 generic.go:334] "Generic (PLEG): container finished" podID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerID="a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2" exitCode=143 Feb 28 10:57:12 crc kubenswrapper[4972]: I0228 10:57:12.657796 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerDied","Data":"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2"} Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.634343 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.688639 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs\") pod \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.688747 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5kks\" (UniqueName: \"kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks\") pod \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.689551 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs" (OuterVolumeSpecName: "logs") pod "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" (UID: "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.690272 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data\") pod \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.690416 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle\") pod \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\" (UID: \"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2\") " Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.693788 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.715807 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks" (OuterVolumeSpecName: "kube-api-access-t5kks") pod "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" (UID: "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2"). InnerVolumeSpecName "kube-api-access-t5kks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.719911 4972 generic.go:334] "Generic (PLEG): container finished" podID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerID="78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec" exitCode=0 Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.719980 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerDied","Data":"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec"} Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.720023 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8aabf8aa-17d0-4e2a-ae12-d667e5545ca2","Type":"ContainerDied","Data":"42caf911a7c67242494c06cd9ece6a39ff0991e4a7b5bb5ccdb8b251800cae29"} Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.720080 4972 scope.go:117] "RemoveContainer" containerID="78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.720083 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.749864 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data" (OuterVolumeSpecName: "config-data") pod "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" (UID: "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.788454 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" (UID: "8aabf8aa-17d0-4e2a-ae12-d667e5545ca2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.795374 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5kks\" (UniqueName: \"kubernetes.io/projected/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-kube-api-access-t5kks\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.795621 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.795725 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.841106 4972 scope.go:117] "RemoveContainer" containerID="a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.872184 4972 scope.go:117] "RemoveContainer" containerID="78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec" Feb 28 10:57:15 crc kubenswrapper[4972]: E0228 10:57:15.872997 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec\": container with ID starting with 78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec not found: ID does not exist" containerID="78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.873091 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec"} err="failed to get container status \"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec\": rpc error: code = NotFound desc = could not find container \"78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec\": container with ID starting with 78e9e87ea6163287ff22fb9c8808d0e25a8a22f7c10d4646b0b8ef75e02a5aec not found: ID does not exist" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.873140 4972 scope.go:117] "RemoveContainer" containerID="a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2" Feb 28 10:57:15 crc kubenswrapper[4972]: E0228 10:57:15.874777 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2\": container with ID starting with a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2 not found: ID does not exist" containerID="a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2" Feb 28 10:57:15 crc kubenswrapper[4972]: I0228 10:57:15.874812 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2"} err="failed to get container status \"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2\": rpc error: code = NotFound desc = could not find container \"a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2\": container with ID starting with a54a7f8d6c7b4613011c488ae6caf317400ac8897eb1c71c84b56995f22d88b2 not found: ID does not exist" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.049547 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.059970 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.090126 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:16 crc kubenswrapper[4972]: E0228 10:57:16.090532 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-api" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.090551 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-api" Feb 28 10:57:16 crc kubenswrapper[4972]: E0228 10:57:16.090565 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-log" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.090572 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-log" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.090762 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-log" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.090790 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" containerName="nova-api-api" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.091714 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.094651 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.097651 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.097651 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.142294 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.203545 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9488h\" (UniqueName: \"kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.203680 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.203797 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.204084 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.204151 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.204343 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.306943 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307033 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307092 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307183 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9488h\" (UniqueName: \"kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307208 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307260 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.307398 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.314235 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.314864 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.318954 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.327561 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.328739 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9488h\" (UniqueName: \"kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h\") pod \"nova-api-0\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.408707 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:16 crc kubenswrapper[4972]: I0228 10:57:16.849172 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:16 crc kubenswrapper[4972]: W0228 10:57:16.921970 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f1a928b_3511_46d3_8904_9f74b52dc952.slice/crio-99200ae31cb94d38d7fcb27e0dc034a072bd25b0ff01d20a2fedecd3e4f90963 WatchSource:0}: Error finding container 99200ae31cb94d38d7fcb27e0dc034a072bd25b0ff01d20a2fedecd3e4f90963: Status 404 returned error can't find the container with id 99200ae31cb94d38d7fcb27e0dc034a072bd25b0ff01d20a2fedecd3e4f90963 Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.166849 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226541 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226695 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226718 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226762 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226840 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptphw\" (UniqueName: \"kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226891 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226911 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.226959 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd\") pod \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\" (UID: \"99923c8e-fce0-44fb-b2b1-ef432ceefe7d\") " Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.228433 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.229270 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.232921 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts" (OuterVolumeSpecName: "scripts") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.233613 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw" (OuterVolumeSpecName: "kube-api-access-ptphw") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "kube-api-access-ptphw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.246221 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.263055 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.275657 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.309030 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330678 4972 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330711 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptphw\" (UniqueName: \"kubernetes.io/projected/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-kube-api-access-ptphw\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330761 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330772 4972 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330780 4972 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.330791 4972 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.343555 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.365054 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data" (OuterVolumeSpecName: "config-data") pod "99923c8e-fce0-44fb-b2b1-ef432ceefe7d" (UID: "99923c8e-fce0-44fb-b2b1-ef432ceefe7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.432316 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.432359 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99923c8e-fce0-44fb-b2b1-ef432ceefe7d-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.783935 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerStarted","Data":"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9"} Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.784336 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerStarted","Data":"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0"} Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.784351 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerStarted","Data":"99200ae31cb94d38d7fcb27e0dc034a072bd25b0ff01d20a2fedecd3e4f90963"} Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.796036 4972 generic.go:334] "Generic (PLEG): container finished" podID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerID="7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8" exitCode=0 Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.796447 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.807085 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aabf8aa-17d0-4e2a-ae12-d667e5545ca2" path="/var/lib/kubelet/pods/8aabf8aa-17d0-4e2a-ae12-d667e5545ca2/volumes" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.808581 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerDied","Data":"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8"} Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.808626 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99923c8e-fce0-44fb-b2b1-ef432ceefe7d","Type":"ContainerDied","Data":"a477d5e7f5dbec9666a8828b4d6bd2d8eade0404b16e204082109d3c1b48a87a"} Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.808657 4972 scope.go:117] "RemoveContainer" containerID="b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.815129 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.815108224 podStartE2EDuration="1.815108224s" podCreationTimestamp="2026-02-28 10:57:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:17.80566949 +0000 UTC m=+1294.717677228" watchObservedRunningTime="2026-02-28 10:57:17.815108224 +0000 UTC m=+1294.727115962" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.835600 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.844863 4972 scope.go:117] "RemoveContainer" containerID="2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.864239 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.893044 4972 scope.go:117] "RemoveContainer" containerID="7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.912786 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.959642 4972 scope.go:117] "RemoveContainer" containerID="101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.998960 4972 scope.go:117] "RemoveContainer" containerID="b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66" Feb 28 10:57:17 crc kubenswrapper[4972]: E0228 10:57:17.999909 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66\": container with ID starting with b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66 not found: ID does not exist" containerID="b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.999945 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66"} err="failed to get container status \"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66\": rpc error: code = NotFound desc = could not find container \"b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66\": container with ID starting with b08894bb90929718496729dc63dcb0a03a926ccd83fbae2ff8f514ade0e10f66 not found: ID does not exist" Feb 28 10:57:17 crc kubenswrapper[4972]: I0228 10:57:17.999966 4972 scope.go:117] "RemoveContainer" containerID="2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.000330 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e\": container with ID starting with 2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e not found: ID does not exist" containerID="2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.000351 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e"} err="failed to get container status \"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e\": rpc error: code = NotFound desc = could not find container \"2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e\": container with ID starting with 2a65bc3068a4b13c075dce4a168df1abd52e0ef2f4623d8708252ba3ed062d1e not found: ID does not exist" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.000365 4972 scope.go:117] "RemoveContainer" containerID="7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.000609 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8\": container with ID starting with 7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8 not found: ID does not exist" containerID="7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.000632 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8"} err="failed to get container status \"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8\": rpc error: code = NotFound desc = could not find container \"7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8\": container with ID starting with 7d19384003d361a22fb360c076c61b8a94dc884ed6d1115dafb2f840edfd03b8 not found: ID does not exist" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.000645 4972 scope.go:117] "RemoveContainer" containerID="101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.000830 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9\": container with ID starting with 101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9 not found: ID does not exist" containerID="101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.000855 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9"} err="failed to get container status \"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9\": rpc error: code = NotFound desc = could not find container \"101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9\": container with ID starting with 101851fcc6962a351f48c8a379fa08fc3b6519cc765e7229438751486cb84af9 not found: ID does not exist" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.002503 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.002921 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="proxy-httpd" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.002940 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="proxy-httpd" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.002954 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-notification-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.002961 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-notification-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.002977 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-central-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.002983 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-central-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: E0228 10:57:18.002991 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="sg-core" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.002996 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="sg-core" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.003221 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-notification-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.003247 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="sg-core" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.003264 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="ceilometer-central-agent" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.003279 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" containerName="proxy-httpd" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.005141 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.012225 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.012243 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.012479 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.019846 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051343 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051426 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-log-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051551 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl76s\" (UniqueName: \"kubernetes.io/projected/0a6e6077-551d-4ef1-b669-3c5bac893957-kube-api-access-cl76s\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051584 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051614 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051655 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-run-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051772 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-scripts\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.051828 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-config-data\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.092868 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8v846"] Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.095021 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.097129 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.098035 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.114843 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8v846"] Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.153866 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-run-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.153948 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-scripts\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.153989 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154007 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-config-data\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154057 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154084 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154144 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-log-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154184 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl76s\" (UniqueName: \"kubernetes.io/projected/0a6e6077-551d-4ef1-b669-3c5bac893957-kube-api-access-cl76s\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154259 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154304 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgll6\" (UniqueName: \"kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.154326 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.155437 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-run-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.156143 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a6e6077-551d-4ef1-b669-3c5bac893957-log-httpd\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.161714 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-scripts\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.166713 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-config-data\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.169051 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.169668 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.179785 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a6e6077-551d-4ef1-b669-3c5bac893957-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.180887 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl76s\" (UniqueName: \"kubernetes.io/projected/0a6e6077-551d-4ef1-b669-3c5bac893957-kube-api-access-cl76s\") pod \"ceilometer-0\" (UID: \"0a6e6077-551d-4ef1-b669-3c5bac893957\") " pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.256465 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.256609 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.257543 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgll6\" (UniqueName: \"kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.257586 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.261503 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.262028 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.262701 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.278201 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgll6\" (UniqueName: \"kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6\") pod \"nova-cell1-cell-mapping-8v846\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.336321 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.411221 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.806266 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 28 10:57:18 crc kubenswrapper[4972]: W0228 10:57:18.815255 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a6e6077_551d_4ef1_b669_3c5bac893957.slice/crio-45bd96ee648b3bc803e62f792de7b9061fa54c0ec8e448e9ed380de3664d67af WatchSource:0}: Error finding container 45bd96ee648b3bc803e62f792de7b9061fa54c0ec8e448e9ed380de3664d67af: Status 404 returned error can't find the container with id 45bd96ee648b3bc803e62f792de7b9061fa54c0ec8e448e9ed380de3664d67af Feb 28 10:57:18 crc kubenswrapper[4972]: W0228 10:57:18.916475 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6537ac03_6028_4248_8e38_93ade43d502f.slice/crio-75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92 WatchSource:0}: Error finding container 75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92: Status 404 returned error can't find the container with id 75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92 Feb 28 10:57:18 crc kubenswrapper[4972]: I0228 10:57:18.920934 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8v846"] Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.465685 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.546331 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.546608 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="dnsmasq-dns" containerID="cri-o://5f5288931eab8244b943d57472b2fecc0ec7c4254595258eba9a03a7acb4da8e" gracePeriod=10 Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.805184 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99923c8e-fce0-44fb-b2b1-ef432ceefe7d" path="/var/lib/kubelet/pods/99923c8e-fce0-44fb-b2b1-ef432ceefe7d/volumes" Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.847879 4972 generic.go:334] "Generic (PLEG): container finished" podID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerID="5f5288931eab8244b943d57472b2fecc0ec7c4254595258eba9a03a7acb4da8e" exitCode=0 Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.848003 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" event={"ID":"8212c91a-9a0e-4873-b7ca-2547d5aa515a","Type":"ContainerDied","Data":"5f5288931eab8244b943d57472b2fecc0ec7c4254595258eba9a03a7acb4da8e"} Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.855437 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8v846" event={"ID":"6537ac03-6028-4248-8e38-93ade43d502f","Type":"ContainerStarted","Data":"8dcd5a9c060a7077f0e6e00d7a468ccc5176b9666930904d50d93fc122d0a42d"} Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.855541 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8v846" event={"ID":"6537ac03-6028-4248-8e38-93ade43d502f","Type":"ContainerStarted","Data":"75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92"} Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.858909 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a6e6077-551d-4ef1-b669-3c5bac893957","Type":"ContainerStarted","Data":"49bdab6fb9683900630e0803f61cccd6d22142250eaa0c7b85ea5914b171c7ee"} Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.858956 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a6e6077-551d-4ef1-b669-3c5bac893957","Type":"ContainerStarted","Data":"45bd96ee648b3bc803e62f792de7b9061fa54c0ec8e448e9ed380de3664d67af"} Feb 28 10:57:19 crc kubenswrapper[4972]: I0228 10:57:19.889571 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8v846" podStartSLOduration=1.8895450870000001 podStartE2EDuration="1.889545087s" podCreationTimestamp="2026-02-28 10:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:19.880192276 +0000 UTC m=+1296.792200014" watchObservedRunningTime="2026-02-28 10:57:19.889545087 +0000 UTC m=+1296.801552825" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.123036 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.306912 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.307434 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.307550 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.308028 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.308087 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk8vp\" (UniqueName: \"kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.308260 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config\") pod \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\" (UID: \"8212c91a-9a0e-4873-b7ca-2547d5aa515a\") " Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.333841 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp" (OuterVolumeSpecName: "kube-api-access-tk8vp") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "kube-api-access-tk8vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.379258 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.386559 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.393212 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config" (OuterVolumeSpecName: "config") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.397197 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.414644 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.414737 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk8vp\" (UniqueName: \"kubernetes.io/projected/8212c91a-9a0e-4873-b7ca-2547d5aa515a-kube-api-access-tk8vp\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.414759 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.414785 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.414799 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.448933 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8212c91a-9a0e-4873-b7ca-2547d5aa515a" (UID: "8212c91a-9a0e-4873-b7ca-2547d5aa515a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.516954 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8212c91a-9a0e-4873-b7ca-2547d5aa515a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.880351 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" event={"ID":"8212c91a-9a0e-4873-b7ca-2547d5aa515a","Type":"ContainerDied","Data":"e3eb544d61f0feee6f07810eeab22304833e86f40745b44950af3c86a1e4abaa"} Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.880423 4972 scope.go:117] "RemoveContainer" containerID="5f5288931eab8244b943d57472b2fecc0ec7c4254595258eba9a03a7acb4da8e" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.880463 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.889444 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a6e6077-551d-4ef1-b669-3c5bac893957","Type":"ContainerStarted","Data":"73b43573f373d030ed9908c6fdc8db5a1bdf168efe9276d2772fcf87f155a04b"} Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.889588 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a6e6077-551d-4ef1-b669-3c5bac893957","Type":"ContainerStarted","Data":"2afafdbc93e407c6185753a72aa8f4478b8d332c3a99ce395f84e82477e06b00"} Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.918711 4972 scope.go:117] "RemoveContainer" containerID="28f5ab92e31d20609202fb169287845d7d75861f5eac59b57889b9984222140c" Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.929534 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:57:20 crc kubenswrapper[4972]: I0228 10:57:20.939918 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwt5h"] Feb 28 10:57:21 crc kubenswrapper[4972]: I0228 10:57:21.802567 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" path="/var/lib/kubelet/pods/8212c91a-9a0e-4873-b7ca-2547d5aa515a/volumes" Feb 28 10:57:22 crc kubenswrapper[4972]: I0228 10:57:22.913475 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a6e6077-551d-4ef1-b669-3c5bac893957","Type":"ContainerStarted","Data":"2731ddb0e193663425857abfb3fb297c10679add99234e43dd56bec31dec3879"} Feb 28 10:57:22 crc kubenswrapper[4972]: I0228 10:57:22.913954 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 28 10:57:22 crc kubenswrapper[4972]: I0228 10:57:22.943687 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.379978877 podStartE2EDuration="5.943669501s" podCreationTimestamp="2026-02-28 10:57:17 +0000 UTC" firstStartedPulling="2026-02-28 10:57:18.817944841 +0000 UTC m=+1295.729952579" lastFinishedPulling="2026-02-28 10:57:22.381635465 +0000 UTC m=+1299.293643203" observedRunningTime="2026-02-28 10:57:22.93718123 +0000 UTC m=+1299.849188978" watchObservedRunningTime="2026-02-28 10:57:22.943669501 +0000 UTC m=+1299.855677239" Feb 28 10:57:24 crc kubenswrapper[4972]: I0228 10:57:24.935897 4972 generic.go:334] "Generic (PLEG): container finished" podID="6537ac03-6028-4248-8e38-93ade43d502f" containerID="8dcd5a9c060a7077f0e6e00d7a468ccc5176b9666930904d50d93fc122d0a42d" exitCode=0 Feb 28 10:57:24 crc kubenswrapper[4972]: I0228 10:57:24.936039 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8v846" event={"ID":"6537ac03-6028-4248-8e38-93ade43d502f","Type":"ContainerDied","Data":"8dcd5a9c060a7077f0e6e00d7a468ccc5176b9666930904d50d93fc122d0a42d"} Feb 28 10:57:25 crc kubenswrapper[4972]: I0228 10:57:25.027201 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-jwt5h" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: i/o timeout" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.395572 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.409536 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.409809 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.549974 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data\") pod \"6537ac03-6028-4248-8e38-93ade43d502f\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.550146 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts\") pod \"6537ac03-6028-4248-8e38-93ade43d502f\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.550286 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle\") pod \"6537ac03-6028-4248-8e38-93ade43d502f\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.550453 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgll6\" (UniqueName: \"kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6\") pod \"6537ac03-6028-4248-8e38-93ade43d502f\" (UID: \"6537ac03-6028-4248-8e38-93ade43d502f\") " Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.555663 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts" (OuterVolumeSpecName: "scripts") pod "6537ac03-6028-4248-8e38-93ade43d502f" (UID: "6537ac03-6028-4248-8e38-93ade43d502f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.571658 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6" (OuterVolumeSpecName: "kube-api-access-bgll6") pod "6537ac03-6028-4248-8e38-93ade43d502f" (UID: "6537ac03-6028-4248-8e38-93ade43d502f"). InnerVolumeSpecName "kube-api-access-bgll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.577906 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6537ac03-6028-4248-8e38-93ade43d502f" (UID: "6537ac03-6028-4248-8e38-93ade43d502f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.598748 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data" (OuterVolumeSpecName: "config-data") pod "6537ac03-6028-4248-8e38-93ade43d502f" (UID: "6537ac03-6028-4248-8e38-93ade43d502f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.655476 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgll6\" (UniqueName: \"kubernetes.io/projected/6537ac03-6028-4248-8e38-93ade43d502f-kube-api-access-bgll6\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.655515 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.655528 4972 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-scripts\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.655540 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6537ac03-6028-4248-8e38-93ade43d502f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.972662 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8v846" Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.972702 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8v846" event={"ID":"6537ac03-6028-4248-8e38-93ade43d502f","Type":"ContainerDied","Data":"75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92"} Feb 28 10:57:26 crc kubenswrapper[4972]: I0228 10:57:26.972739 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75e26f9b0e0493a4fe4c4d7ba71adf4c9780f8061d7a158cafcfee445c55db92" Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.166329 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.179330 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.179699 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="18465c59-c525-48fd-8898-3ae76bd54d11" containerName="nova-scheduler-scheduler" containerID="cri-o://28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04" gracePeriod=30 Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.260698 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.260938 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" containerID="cri-o://f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a" gracePeriod=30 Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.261037 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" containerID="cri-o://b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024" gracePeriod=30 Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.420694 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.420742 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.989720 4972 generic.go:334] "Generic (PLEG): container finished" podID="a06a1db6-366f-4838-93ec-a749eae64f71" containerID="f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a" exitCode=143 Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.989804 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerDied","Data":"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a"} Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.990434 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-log" containerID="cri-o://6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0" gracePeriod=30 Feb 28 10:57:27 crc kubenswrapper[4972]: I0228 10:57:27.990429 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-api" containerID="cri-o://34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9" gracePeriod=30 Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.402358 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.487839 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data\") pod \"18465c59-c525-48fd-8898-3ae76bd54d11\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.487975 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl6rj\" (UniqueName: \"kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj\") pod \"18465c59-c525-48fd-8898-3ae76bd54d11\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.488357 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle\") pod \"18465c59-c525-48fd-8898-3ae76bd54d11\" (UID: \"18465c59-c525-48fd-8898-3ae76bd54d11\") " Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.494589 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj" (OuterVolumeSpecName: "kube-api-access-kl6rj") pod "18465c59-c525-48fd-8898-3ae76bd54d11" (UID: "18465c59-c525-48fd-8898-3ae76bd54d11"). InnerVolumeSpecName "kube-api-access-kl6rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.518410 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18465c59-c525-48fd-8898-3ae76bd54d11" (UID: "18465c59-c525-48fd-8898-3ae76bd54d11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.521894 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data" (OuterVolumeSpecName: "config-data") pod "18465c59-c525-48fd-8898-3ae76bd54d11" (UID: "18465c59-c525-48fd-8898-3ae76bd54d11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.592046 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl6rj\" (UniqueName: \"kubernetes.io/projected/18465c59-c525-48fd-8898-3ae76bd54d11-kube-api-access-kl6rj\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.592112 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.592123 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18465c59-c525-48fd-8898-3ae76bd54d11-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.890909 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:57:28 crc kubenswrapper[4972]: I0228 10:57:28.891284 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.005137 4972 generic.go:334] "Generic (PLEG): container finished" podID="18465c59-c525-48fd-8898-3ae76bd54d11" containerID="28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04" exitCode=0 Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.005221 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.005274 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18465c59-c525-48fd-8898-3ae76bd54d11","Type":"ContainerDied","Data":"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04"} Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.005327 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18465c59-c525-48fd-8898-3ae76bd54d11","Type":"ContainerDied","Data":"7d14d47232d853f70f13b97f0d2b626a4ca22d04bbcfb69ae01460a8b3be0204"} Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.005355 4972 scope.go:117] "RemoveContainer" containerID="28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.012337 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerID="6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0" exitCode=143 Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.012398 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerDied","Data":"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0"} Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.041924 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.052337 4972 scope.go:117] "RemoveContainer" containerID="28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04" Feb 28 10:57:29 crc kubenswrapper[4972]: E0228 10:57:29.053066 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04\": container with ID starting with 28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04 not found: ID does not exist" containerID="28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.053135 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04"} err="failed to get container status \"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04\": rpc error: code = NotFound desc = could not find container \"28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04\": container with ID starting with 28241661d78e59225aab9345aac2d389b289c9d02870daf222aa78897c24de04 not found: ID does not exist" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.067162 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.075308 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:29 crc kubenswrapper[4972]: E0228 10:57:29.076095 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="dnsmasq-dns" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.076186 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="dnsmasq-dns" Feb 28 10:57:29 crc kubenswrapper[4972]: E0228 10:57:29.076267 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="init" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.076334 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="init" Feb 28 10:57:29 crc kubenswrapper[4972]: E0228 10:57:29.076402 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6537ac03-6028-4248-8e38-93ade43d502f" containerName="nova-manage" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.076482 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="6537ac03-6028-4248-8e38-93ade43d502f" containerName="nova-manage" Feb 28 10:57:29 crc kubenswrapper[4972]: E0228 10:57:29.076561 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18465c59-c525-48fd-8898-3ae76bd54d11" containerName="nova-scheduler-scheduler" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.076619 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="18465c59-c525-48fd-8898-3ae76bd54d11" containerName="nova-scheduler-scheduler" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.076978 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="6537ac03-6028-4248-8e38-93ade43d502f" containerName="nova-manage" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.077057 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8212c91a-9a0e-4873-b7ca-2547d5aa515a" containerName="dnsmasq-dns" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.077198 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="18465c59-c525-48fd-8898-3ae76bd54d11" containerName="nova-scheduler-scheduler" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.077960 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.082493 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.085758 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.201900 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-config-data\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.202389 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sflw\" (UniqueName: \"kubernetes.io/projected/44f084cb-b672-497c-8c66-23785c9971f7-kube-api-access-4sflw\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.202522 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.304805 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sflw\" (UniqueName: \"kubernetes.io/projected/44f084cb-b672-497c-8c66-23785c9971f7-kube-api-access-4sflw\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.304870 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.304984 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-config-data\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.312889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-config-data\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.314672 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f084cb-b672-497c-8c66-23785c9971f7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.326864 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sflw\" (UniqueName: \"kubernetes.io/projected/44f084cb-b672-497c-8c66-23785c9971f7-kube-api-access-4sflw\") pod \"nova-scheduler-0\" (UID: \"44f084cb-b672-497c-8c66-23785c9971f7\") " pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.399920 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.801159 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18465c59-c525-48fd-8898-3ae76bd54d11" path="/var/lib/kubelet/pods/18465c59-c525-48fd-8898-3ae76bd54d11/volumes" Feb 28 10:57:29 crc kubenswrapper[4972]: I0228 10:57:29.905825 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 28 10:57:29 crc kubenswrapper[4972]: W0228 10:57:29.917720 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44f084cb_b672_497c_8c66_23785c9971f7.slice/crio-d46407808668ca9d9dc01b61539cfeedeac98785d681259a08af315531400bec WatchSource:0}: Error finding container d46407808668ca9d9dc01b61539cfeedeac98785d681259a08af315531400bec: Status 404 returned error can't find the container with id d46407808668ca9d9dc01b61539cfeedeac98785d681259a08af315531400bec Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.025293 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44f084cb-b672-497c-8c66-23785c9971f7","Type":"ContainerStarted","Data":"d46407808668ca9d9dc01b61539cfeedeac98785d681259a08af315531400bec"} Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.390069 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:58724->10.217.0.200:8775: read: connection reset by peer" Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.390138 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:58726->10.217.0.200:8775: read: connection reset by peer" Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.870972 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.956481 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs\") pod \"a06a1db6-366f-4838-93ec-a749eae64f71\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.956549 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs\") pod \"a06a1db6-366f-4838-93ec-a749eae64f71\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.956584 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w29d\" (UniqueName: \"kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d\") pod \"a06a1db6-366f-4838-93ec-a749eae64f71\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.956676 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data\") pod \"a06a1db6-366f-4838-93ec-a749eae64f71\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.956693 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle\") pod \"a06a1db6-366f-4838-93ec-a749eae64f71\" (UID: \"a06a1db6-366f-4838-93ec-a749eae64f71\") " Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.957421 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs" (OuterVolumeSpecName: "logs") pod "a06a1db6-366f-4838-93ec-a749eae64f71" (UID: "a06a1db6-366f-4838-93ec-a749eae64f71"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.971247 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d" (OuterVolumeSpecName: "kube-api-access-7w29d") pod "a06a1db6-366f-4838-93ec-a749eae64f71" (UID: "a06a1db6-366f-4838-93ec-a749eae64f71"). InnerVolumeSpecName "kube-api-access-7w29d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:30 crc kubenswrapper[4972]: I0228 10:57:30.991239 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data" (OuterVolumeSpecName: "config-data") pod "a06a1db6-366f-4838-93ec-a749eae64f71" (UID: "a06a1db6-366f-4838-93ec-a749eae64f71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.007651 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a06a1db6-366f-4838-93ec-a749eae64f71" (UID: "a06a1db6-366f-4838-93ec-a749eae64f71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.023037 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a06a1db6-366f-4838-93ec-a749eae64f71" (UID: "a06a1db6-366f-4838-93ec-a749eae64f71"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.041305 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"44f084cb-b672-497c-8c66-23785c9971f7","Type":"ContainerStarted","Data":"98a7e20ca59e0e28629e5ae847fff8069da87063d344ec116b25ba1accae19c9"} Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.048264 4972 generic.go:334] "Generic (PLEG): container finished" podID="a06a1db6-366f-4838-93ec-a749eae64f71" containerID="b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024" exitCode=0 Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.048338 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.048319 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerDied","Data":"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024"} Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.048563 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a06a1db6-366f-4838-93ec-a749eae64f71","Type":"ContainerDied","Data":"f66115d1a169dff011b58bccba8987b74bb84f49d3a51204bd68d4d70b845fd1"} Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.048616 4972 scope.go:117] "RemoveContainer" containerID="b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.059729 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a06a1db6-366f-4838-93ec-a749eae64f71-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.059937 4972 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.060019 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w29d\" (UniqueName: \"kubernetes.io/projected/a06a1db6-366f-4838-93ec-a749eae64f71-kube-api-access-7w29d\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.060078 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.060136 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a06a1db6-366f-4838-93ec-a749eae64f71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.068806 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.068791305 podStartE2EDuration="2.068791305s" podCreationTimestamp="2026-02-28 10:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:31.057082928 +0000 UTC m=+1307.969090666" watchObservedRunningTime="2026-02-28 10:57:31.068791305 +0000 UTC m=+1307.980799043" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.139593 4972 scope.go:117] "RemoveContainer" containerID="f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.155159 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.167669 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.175710 4972 scope.go:117] "RemoveContainer" containerID="b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024" Feb 28 10:57:31 crc kubenswrapper[4972]: E0228 10:57:31.176129 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024\": container with ID starting with b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024 not found: ID does not exist" containerID="b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.176166 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024"} err="failed to get container status \"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024\": rpc error: code = NotFound desc = could not find container \"b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024\": container with ID starting with b12a4d85177e84d3f708d369fedbf81c9217100d6c20f2bf2662cedb92512024 not found: ID does not exist" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.176189 4972 scope.go:117] "RemoveContainer" containerID="f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a" Feb 28 10:57:31 crc kubenswrapper[4972]: E0228 10:57:31.176430 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a\": container with ID starting with f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a not found: ID does not exist" containerID="f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.176452 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a"} err="failed to get container status \"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a\": rpc error: code = NotFound desc = could not find container \"f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a\": container with ID starting with f50690ce87c66a384169b233b845b4ce971a7a23b03d7da234ae1fb1232a429a not found: ID does not exist" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.185566 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:31 crc kubenswrapper[4972]: E0228 10:57:31.186365 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.186386 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" Feb 28 10:57:31 crc kubenswrapper[4972]: E0228 10:57:31.186438 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.186449 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.186754 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-log" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.186788 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" containerName="nova-metadata-metadata" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.188286 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.192220 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.192813 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.203441 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.264207 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-config-data\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.264624 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92b274e8-6674-42a1-81f3-302d28e22926-logs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.265122 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.265390 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmdfp\" (UniqueName: \"kubernetes.io/projected/92b274e8-6674-42a1-81f3-302d28e22926-kube-api-access-jmdfp\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.265697 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.367693 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmdfp\" (UniqueName: \"kubernetes.io/projected/92b274e8-6674-42a1-81f3-302d28e22926-kube-api-access-jmdfp\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.367762 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.367828 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-config-data\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.367868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92b274e8-6674-42a1-81f3-302d28e22926-logs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.367923 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.368891 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92b274e8-6674-42a1-81f3-302d28e22926-logs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.373074 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-config-data\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.373432 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.373789 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/92b274e8-6674-42a1-81f3-302d28e22926-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.392267 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmdfp\" (UniqueName: \"kubernetes.io/projected/92b274e8-6674-42a1-81f3-302d28e22926-kube-api-access-jmdfp\") pod \"nova-metadata-0\" (UID: \"92b274e8-6674-42a1-81f3-302d28e22926\") " pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.539928 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 28 10:57:31 crc kubenswrapper[4972]: I0228 10:57:31.805516 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06a1db6-366f-4838-93ec-a749eae64f71" path="/var/lib/kubelet/pods/a06a1db6-366f-4838-93ec-a749eae64f71/volumes" Feb 28 10:57:32 crc kubenswrapper[4972]: I0228 10:57:32.017613 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 28 10:57:32 crc kubenswrapper[4972]: W0228 10:57:32.027263 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92b274e8_6674_42a1_81f3_302d28e22926.slice/crio-25684de5b2a512468a1818505ee15ec2795f9936150fe739b60bcbfeac06982a WatchSource:0}: Error finding container 25684de5b2a512468a1818505ee15ec2795f9936150fe739b60bcbfeac06982a: Status 404 returned error can't find the container with id 25684de5b2a512468a1818505ee15ec2795f9936150fe739b60bcbfeac06982a Feb 28 10:57:32 crc kubenswrapper[4972]: I0228 10:57:32.062035 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92b274e8-6674-42a1-81f3-302d28e22926","Type":"ContainerStarted","Data":"25684de5b2a512468a1818505ee15ec2795f9936150fe739b60bcbfeac06982a"} Feb 28 10:57:32 crc kubenswrapper[4972]: I0228 10:57:32.961909 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.006733 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.007590 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.007679 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.007865 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.007941 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9488h\" (UniqueName: \"kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.008156 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs\") pod \"9f1a928b-3511-46d3-8904-9f74b52dc952\" (UID: \"9f1a928b-3511-46d3-8904-9f74b52dc952\") " Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.010017 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs" (OuterVolumeSpecName: "logs") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.010294 4972 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f1a928b-3511-46d3-8904-9f74b52dc952-logs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.014603 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h" (OuterVolumeSpecName: "kube-api-access-9488h") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "kube-api-access-9488h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.038172 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data" (OuterVolumeSpecName: "config-data") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.048841 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.074319 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.081515 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92b274e8-6674-42a1-81f3-302d28e22926","Type":"ContainerStarted","Data":"fb2f1a792fa60e434b6e5cf0720913289f7031e585ea52505e1e90bac8c1a08f"} Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.081571 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92b274e8-6674-42a1-81f3-302d28e22926","Type":"ContainerStarted","Data":"e03a09141e074c6c06f6219a0f8dcc5c783aad149d02f77c304604139fe29cd5"} Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.086768 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerID="34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9" exitCode=0 Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.086839 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerDied","Data":"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9"} Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.086887 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f1a928b-3511-46d3-8904-9f74b52dc952","Type":"ContainerDied","Data":"99200ae31cb94d38d7fcb27e0dc034a072bd25b0ff01d20a2fedecd3e4f90963"} Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.086922 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.086941 4972 scope.go:117] "RemoveContainer" containerID="34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.090241 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9f1a928b-3511-46d3-8904-9f74b52dc952" (UID: "9f1a928b-3511-46d3-8904-9f74b52dc952"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.113525 4972 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.113570 4972 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.113590 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.113604 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f1a928b-3511-46d3-8904-9f74b52dc952-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.113617 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9488h\" (UniqueName: \"kubernetes.io/projected/9f1a928b-3511-46d3-8904-9f74b52dc952-kube-api-access-9488h\") on node \"crc\" DevicePath \"\"" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.117128 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.117008716 podStartE2EDuration="2.117008716s" podCreationTimestamp="2026-02-28 10:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:33.107053038 +0000 UTC m=+1310.019060776" watchObservedRunningTime="2026-02-28 10:57:33.117008716 +0000 UTC m=+1310.029016454" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.131078 4972 scope.go:117] "RemoveContainer" containerID="6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.153633 4972 scope.go:117] "RemoveContainer" containerID="34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9" Feb 28 10:57:33 crc kubenswrapper[4972]: E0228 10:57:33.154144 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9\": container with ID starting with 34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9 not found: ID does not exist" containerID="34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.154193 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9"} err="failed to get container status \"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9\": rpc error: code = NotFound desc = could not find container \"34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9\": container with ID starting with 34d776ba5ee3d453da2fcc6549b0982a39c9900229feaea2802c909c26434de9 not found: ID does not exist" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.154222 4972 scope.go:117] "RemoveContainer" containerID="6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0" Feb 28 10:57:33 crc kubenswrapper[4972]: E0228 10:57:33.154698 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0\": container with ID starting with 6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0 not found: ID does not exist" containerID="6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.154763 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0"} err="failed to get container status \"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0\": rpc error: code = NotFound desc = could not find container \"6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0\": container with ID starting with 6b8e575f909f0c8f2fd913d4db8f9936bbebd631a13a5225c94543f48ebf4eb0 not found: ID does not exist" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.459706 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.475954 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.487391 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:33 crc kubenswrapper[4972]: E0228 10:57:33.487826 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-log" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.487841 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-log" Feb 28 10:57:33 crc kubenswrapper[4972]: E0228 10:57:33.487861 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-api" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.487867 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-api" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.488120 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-log" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.488140 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" containerName="nova-api-api" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.489227 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.496539 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.497331 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.497787 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.504520 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.629262 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.629336 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-config-data\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.629383 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.629518 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64af5818-baad-43d6-abb6-644fb18b3b5e-logs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.629620 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.630138 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgqww\" (UniqueName: \"kubernetes.io/projected/64af5818-baad-43d6-abb6-644fb18b3b5e-kube-api-access-tgqww\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733165 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgqww\" (UniqueName: \"kubernetes.io/projected/64af5818-baad-43d6-abb6-644fb18b3b5e-kube-api-access-tgqww\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733427 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733474 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-config-data\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733516 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733539 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64af5818-baad-43d6-abb6-644fb18b3b5e-logs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.733557 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.734718 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64af5818-baad-43d6-abb6-644fb18b3b5e-logs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.739213 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.739689 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.740604 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-public-tls-certs\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.740806 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64af5818-baad-43d6-abb6-644fb18b3b5e-config-data\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.753669 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgqww\" (UniqueName: \"kubernetes.io/projected/64af5818-baad-43d6-abb6-644fb18b3b5e-kube-api-access-tgqww\") pod \"nova-api-0\" (UID: \"64af5818-baad-43d6-abb6-644fb18b3b5e\") " pod="openstack/nova-api-0" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.800850 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f1a928b-3511-46d3-8904-9f74b52dc952" path="/var/lib/kubelet/pods/9f1a928b-3511-46d3-8904-9f74b52dc952/volumes" Feb 28 10:57:33 crc kubenswrapper[4972]: I0228 10:57:33.827479 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 28 10:57:34 crc kubenswrapper[4972]: W0228 10:57:34.272680 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64af5818_baad_43d6_abb6_644fb18b3b5e.slice/crio-d47928b0030780a89fe05508157718b2bd5e2f220f9478fd91cb83da3637657b WatchSource:0}: Error finding container d47928b0030780a89fe05508157718b2bd5e2f220f9478fd91cb83da3637657b: Status 404 returned error can't find the container with id d47928b0030780a89fe05508157718b2bd5e2f220f9478fd91cb83da3637657b Feb 28 10:57:34 crc kubenswrapper[4972]: I0228 10:57:34.277851 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 28 10:57:34 crc kubenswrapper[4972]: I0228 10:57:34.400915 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 28 10:57:35 crc kubenswrapper[4972]: I0228 10:57:35.115731 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64af5818-baad-43d6-abb6-644fb18b3b5e","Type":"ContainerStarted","Data":"2b61d4b2f7d95babf08d7b6bef87c9aa6095e1d5ca7039b5fe7af00ff50b3ee2"} Feb 28 10:57:35 crc kubenswrapper[4972]: I0228 10:57:35.116032 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64af5818-baad-43d6-abb6-644fb18b3b5e","Type":"ContainerStarted","Data":"01dbbee9d77d2453a299dea24f8e86dadd808584637990cc554292c6203326a4"} Feb 28 10:57:35 crc kubenswrapper[4972]: I0228 10:57:35.116045 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64af5818-baad-43d6-abb6-644fb18b3b5e","Type":"ContainerStarted","Data":"d47928b0030780a89fe05508157718b2bd5e2f220f9478fd91cb83da3637657b"} Feb 28 10:57:35 crc kubenswrapper[4972]: I0228 10:57:35.156069 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.156042071 podStartE2EDuration="2.156042071s" podCreationTimestamp="2026-02-28 10:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:57:35.140053584 +0000 UTC m=+1312.052061322" watchObservedRunningTime="2026-02-28 10:57:35.156042071 +0000 UTC m=+1312.068049829" Feb 28 10:57:36 crc kubenswrapper[4972]: I0228 10:57:36.540324 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 28 10:57:36 crc kubenswrapper[4972]: I0228 10:57:36.540998 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 28 10:57:39 crc kubenswrapper[4972]: I0228 10:57:39.400470 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 28 10:57:39 crc kubenswrapper[4972]: I0228 10:57:39.438081 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 28 10:57:40 crc kubenswrapper[4972]: I0228 10:57:40.199739 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 28 10:57:41 crc kubenswrapper[4972]: I0228 10:57:41.540956 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 28 10:57:41 crc kubenswrapper[4972]: I0228 10:57:41.541707 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 28 10:57:42 crc kubenswrapper[4972]: I0228 10:57:42.554666 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="92b274e8-6674-42a1-81f3-302d28e22926" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:42 crc kubenswrapper[4972]: I0228 10:57:42.554707 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="92b274e8-6674-42a1-81f3-302d28e22926" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:43 crc kubenswrapper[4972]: I0228 10:57:43.827723 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:57:43 crc kubenswrapper[4972]: I0228 10:57:43.828020 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 28 10:57:44 crc kubenswrapper[4972]: I0228 10:57:44.846653 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64af5818-baad-43d6-abb6-644fb18b3b5e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:44 crc kubenswrapper[4972]: I0228 10:57:44.846679 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64af5818-baad-43d6-abb6-644fb18b3b5e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 28 10:57:48 crc kubenswrapper[4972]: I0228 10:57:48.349193 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 28 10:57:51 crc kubenswrapper[4972]: I0228 10:57:51.546622 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 28 10:57:51 crc kubenswrapper[4972]: I0228 10:57:51.548225 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 28 10:57:51 crc kubenswrapper[4972]: I0228 10:57:51.555610 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 28 10:57:51 crc kubenswrapper[4972]: I0228 10:57:51.555698 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.838619 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.839508 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.839932 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.839981 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.847669 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 28 10:57:53 crc kubenswrapper[4972]: I0228 10:57:53.852262 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 28 10:57:58 crc kubenswrapper[4972]: I0228 10:57:58.890760 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:57:58 crc kubenswrapper[4972]: I0228 10:57:58.891063 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.177029 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537938-lhgkx"] Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.194663 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.203366 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.203610 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.203869 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.232340 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537938-lhgkx"] Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.236215 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdlrq\" (UniqueName: \"kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq\") pod \"auto-csr-approver-29537938-lhgkx\" (UID: \"f217308b-0a5b-4723-bf3f-b963474cbdbb\") " pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.338920 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdlrq\" (UniqueName: \"kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq\") pod \"auto-csr-approver-29537938-lhgkx\" (UID: \"f217308b-0a5b-4723-bf3f-b963474cbdbb\") " pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.366744 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdlrq\" (UniqueName: \"kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq\") pod \"auto-csr-approver-29537938-lhgkx\" (UID: \"f217308b-0a5b-4723-bf3f-b963474cbdbb\") " pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:00 crc kubenswrapper[4972]: I0228 10:58:00.534024 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:01 crc kubenswrapper[4972]: I0228 10:58:01.055930 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537938-lhgkx"] Feb 28 10:58:01 crc kubenswrapper[4972]: I0228 10:58:01.159390 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:01 crc kubenswrapper[4972]: I0228 10:58:01.408741 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" event={"ID":"f217308b-0a5b-4723-bf3f-b963474cbdbb","Type":"ContainerStarted","Data":"cffe4f89865becb12816c5921e7b58885c65974ef7023362589936d4b939d0f8"} Feb 28 10:58:02 crc kubenswrapper[4972]: I0228 10:58:02.257159 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:02 crc kubenswrapper[4972]: I0228 10:58:02.419660 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" event={"ID":"f217308b-0a5b-4723-bf3f-b963474cbdbb","Type":"ContainerStarted","Data":"23096be105c71aa1516e708d9e7943de7c044738933eaf0c321ef445e9fd6661"} Feb 28 10:58:02 crc kubenswrapper[4972]: I0228 10:58:02.457217 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" podStartSLOduration=1.635947343 podStartE2EDuration="2.45719632s" podCreationTimestamp="2026-02-28 10:58:00 +0000 UTC" firstStartedPulling="2026-02-28 10:58:01.067732324 +0000 UTC m=+1337.979740082" lastFinishedPulling="2026-02-28 10:58:01.888981321 +0000 UTC m=+1338.800989059" observedRunningTime="2026-02-28 10:58:02.447428248 +0000 UTC m=+1339.359435996" watchObservedRunningTime="2026-02-28 10:58:02.45719632 +0000 UTC m=+1339.369204068" Feb 28 10:58:03 crc kubenswrapper[4972]: I0228 10:58:03.430952 4972 generic.go:334] "Generic (PLEG): container finished" podID="f217308b-0a5b-4723-bf3f-b963474cbdbb" containerID="23096be105c71aa1516e708d9e7943de7c044738933eaf0c321ef445e9fd6661" exitCode=0 Feb 28 10:58:03 crc kubenswrapper[4972]: I0228 10:58:03.431031 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" event={"ID":"f217308b-0a5b-4723-bf3f-b963474cbdbb","Type":"ContainerDied","Data":"23096be105c71aa1516e708d9e7943de7c044738933eaf0c321ef445e9fd6661"} Feb 28 10:58:04 crc kubenswrapper[4972]: I0228 10:58:04.872388 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:04 crc kubenswrapper[4972]: I0228 10:58:04.942147 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdlrq\" (UniqueName: \"kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq\") pod \"f217308b-0a5b-4723-bf3f-b963474cbdbb\" (UID: \"f217308b-0a5b-4723-bf3f-b963474cbdbb\") " Feb 28 10:58:04 crc kubenswrapper[4972]: I0228 10:58:04.950262 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq" (OuterVolumeSpecName: "kube-api-access-tdlrq") pod "f217308b-0a5b-4723-bf3f-b963474cbdbb" (UID: "f217308b-0a5b-4723-bf3f-b963474cbdbb"). InnerVolumeSpecName "kube-api-access-tdlrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.045574 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdlrq\" (UniqueName: \"kubernetes.io/projected/f217308b-0a5b-4723-bf3f-b963474cbdbb-kube-api-access-tdlrq\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.453112 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" event={"ID":"f217308b-0a5b-4723-bf3f-b963474cbdbb","Type":"ContainerDied","Data":"cffe4f89865becb12816c5921e7b58885c65974ef7023362589936d4b939d0f8"} Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.453557 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cffe4f89865becb12816c5921e7b58885c65974ef7023362589936d4b939d0f8" Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.453236 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537938-lhgkx" Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.538623 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537932-5ql5s"] Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.549406 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537932-5ql5s"] Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.802143 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a7b699-f95c-43cd-b919-8ae540010828" path="/var/lib/kubelet/pods/58a7b699-f95c-43cd-b919-8ae540010828/volumes" Feb 28 10:58:05 crc kubenswrapper[4972]: I0228 10:58:05.856800 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="rabbitmq" containerID="cri-o://331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09" gracePeriod=604796 Feb 28 10:58:06 crc kubenswrapper[4972]: I0228 10:58:06.946923 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="rabbitmq" containerID="cri-o://0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731" gracePeriod=604796 Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.458095 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495504 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2w9p\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495580 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495627 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495681 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495725 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495746 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495782 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495836 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495857 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495894 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.495955 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"2d65675a-7f0e-44b3-93df-5aad91061871\" (UID: \"2d65675a-7f0e-44b3-93df-5aad91061871\") " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.496017 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.496333 4972 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.496756 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.497882 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.502245 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info" (OuterVolumeSpecName: "pod-info") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.518081 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.519999 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.520061 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p" (OuterVolumeSpecName: "kube-api-access-f2w9p") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "kube-api-access-f2w9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.522696 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.571534 4972 generic.go:334] "Generic (PLEG): container finished" podID="2d65675a-7f0e-44b3-93df-5aad91061871" containerID="331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09" exitCode=0 Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.571590 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerDied","Data":"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09"} Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.571621 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2d65675a-7f0e-44b3-93df-5aad91061871","Type":"ContainerDied","Data":"7a2426e6cfc347cc1a717c9a966306104dd25d30c2b178309ab403c7304eba33"} Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.571642 4972 scope.go:117] "RemoveContainer" containerID="331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.571836 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.575385 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data" (OuterVolumeSpecName: "config-data") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.598713 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf" (OuterVolumeSpecName: "server-conf") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599777 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599815 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2w9p\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-kube-api-access-f2w9p\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599825 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599833 4972 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d65675a-7f0e-44b3-93df-5aad91061871-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599842 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599850 4972 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-server-conf\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599857 4972 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d65675a-7f0e-44b3-93df-5aad91061871-pod-info\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599866 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.599874 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d65675a-7f0e-44b3-93df-5aad91061871-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.681694 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.701044 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:12 crc kubenswrapper[4972]: E0228 10:58:12.702138 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="setup-container" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.702170 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="setup-container" Feb 28 10:58:12 crc kubenswrapper[4972]: E0228 10:58:12.702211 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="rabbitmq" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.702219 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="rabbitmq" Feb 28 10:58:12 crc kubenswrapper[4972]: E0228 10:58:12.702257 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f217308b-0a5b-4723-bf3f-b963474cbdbb" containerName="oc" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.702266 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f217308b-0a5b-4723-bf3f-b963474cbdbb" containerName="oc" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.702774 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" containerName="rabbitmq" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.702847 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f217308b-0a5b-4723-bf3f-b963474cbdbb" containerName="oc" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.703396 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.705297 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.707400 4972 scope.go:117] "RemoveContainer" containerID="fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.711747 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.722759 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.751977 4972 scope.go:117] "RemoveContainer" containerID="331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09" Feb 28 10:58:12 crc kubenswrapper[4972]: E0228 10:58:12.752431 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09\": container with ID starting with 331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09 not found: ID does not exist" containerID="331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.752612 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09"} err="failed to get container status \"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09\": rpc error: code = NotFound desc = could not find container \"331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09\": container with ID starting with 331b47fffd9dea0b61749fe343125d751f0ef4b08b726bcd9c647f41a72ebe09 not found: ID does not exist" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.752645 4972 scope.go:117] "RemoveContainer" containerID="fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b" Feb 28 10:58:12 crc kubenswrapper[4972]: E0228 10:58:12.754173 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b\": container with ID starting with fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b not found: ID does not exist" containerID="fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.754214 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b"} err="failed to get container status \"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b\": rpc error: code = NotFound desc = could not find container \"fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b\": container with ID starting with fee118c9e3569f5688775bfacead2d49a6eb50a31de9899bfd4943be9bdc7c5b not found: ID does not exist" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.756515 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2d65675a-7f0e-44b3-93df-5aad91061871" (UID: "2d65675a-7f0e-44b3-93df-5aad91061871"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.806433 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.806544 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.806792 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.806851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlq94\" (UniqueName: \"kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.806937 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.807032 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.807118 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.807479 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d65675a-7f0e-44b3-93df-5aad91061871-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910566 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910632 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910673 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910705 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910744 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910775 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlq94\" (UniqueName: \"kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.910812 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.912042 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.912300 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.912314 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.912848 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.913113 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.913265 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.914294 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.923525 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.932254 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlq94\" (UniqueName: \"kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94\") pod \"dnsmasq-dns-79bd4cc8c9-hhpsd\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.936237 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.937778 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.941941 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.942194 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.942412 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jxt7b" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.942709 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.943098 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.943379 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.943649 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 28 10:58:12 crc kubenswrapper[4972]: I0228 10:58:12.971016 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.012896 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.012971 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a821361-b8f7-4511-9f07-9f05f1b8425d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013007 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013024 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lch88\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-kube-api-access-lch88\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013073 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013141 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013169 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013187 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013251 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013270 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.013299 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a821361-b8f7-4511-9f07-9f05f1b8425d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.044641 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116209 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116292 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116334 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116357 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116492 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116523 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116563 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a821361-b8f7-4511-9f07-9f05f1b8425d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116627 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116659 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a821361-b8f7-4511-9f07-9f05f1b8425d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116715 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.116758 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lch88\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-kube-api-access-lch88\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.117316 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.117767 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.117855 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.118159 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.118382 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.119075 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a821361-b8f7-4511-9f07-9f05f1b8425d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.122329 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.138774 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a821361-b8f7-4511-9f07-9f05f1b8425d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.138974 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a821361-b8f7-4511-9f07-9f05f1b8425d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.140754 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.153515 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lch88\" (UniqueName: \"kubernetes.io/projected/6a821361-b8f7-4511-9f07-9f05f1b8425d-kube-api-access-lch88\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.167028 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6a821361-b8f7-4511-9f07-9f05f1b8425d\") " pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.259087 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.501697 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531449 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531674 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxmdw\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531714 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531744 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531773 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531796 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531821 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531892 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.531952 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.532026 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.532061 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins\") pod \"ffd40994-f4d6-442c-97fa-6ed417500617\" (UID: \"ffd40994-f4d6-442c-97fa-6ed417500617\") " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.535536 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.541291 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.545858 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw" (OuterVolumeSpecName: "kube-api-access-fxmdw") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "kube-api-access-fxmdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.545917 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.547588 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.551302 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.551486 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info" (OuterVolumeSpecName: "pod-info") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.556662 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557056 4972 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557094 4972 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ffd40994-f4d6-442c-97fa-6ed417500617-pod-info\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557109 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557121 4972 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ffd40994-f4d6-442c-97fa-6ed417500617-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557133 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557146 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxmdw\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-kube-api-access-fxmdw\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.557197 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.561637 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.589615 4972 generic.go:334] "Generic (PLEG): container finished" podID="ffd40994-f4d6-442c-97fa-6ed417500617" containerID="0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731" exitCode=0 Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.590085 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.590098 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerDied","Data":"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731"} Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.590184 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ffd40994-f4d6-442c-97fa-6ed417500617","Type":"ContainerDied","Data":"aa7446a5a43998a2fbee8e2ee785ebdb631ce8107b26032d4d3066cccc55bdf0"} Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.590247 4972 scope.go:117] "RemoveContainer" containerID="0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.597107 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" event={"ID":"886a175a-2921-4867-9d20-ebc327478bae","Type":"ContainerStarted","Data":"073a8dd505de434f55bdcdcbdd0292512d67a61ec22179ee5b383c50de032a6a"} Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.599495 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data" (OuterVolumeSpecName: "config-data") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.616529 4972 scope.go:117] "RemoveContainer" containerID="072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.634369 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.639288 4972 scope.go:117] "RemoveContainer" containerID="0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731" Feb 28 10:58:13 crc kubenswrapper[4972]: E0228 10:58:13.639837 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731\": container with ID starting with 0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731 not found: ID does not exist" containerID="0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.639870 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731"} err="failed to get container status \"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731\": rpc error: code = NotFound desc = could not find container \"0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731\": container with ID starting with 0619beec9a81a06a0126bcfe0617cc435c970b82cf1f9ec7aa185b37d7ecb731 not found: ID does not exist" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.639890 4972 scope.go:117] "RemoveContainer" containerID="072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a" Feb 28 10:58:13 crc kubenswrapper[4972]: E0228 10:58:13.640210 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a\": container with ID starting with 072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a not found: ID does not exist" containerID="072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.640236 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a"} err="failed to get container status \"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a\": rpc error: code = NotFound desc = could not find container \"072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a\": container with ID starting with 072ff5707bca20ceedcec66cd947de996dd8c0c14f76a81416b5bd036dc2b00a not found: ID does not exist" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.645798 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf" (OuterVolumeSpecName: "server-conf") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.658625 4972 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-server-conf\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.658662 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.658672 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ffd40994-f4d6-442c-97fa-6ed417500617-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.658681 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.702041 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ffd40994-f4d6-442c-97fa-6ed417500617" (UID: "ffd40994-f4d6-442c-97fa-6ed417500617"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.761748 4972 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ffd40994-f4d6-442c-97fa-6ed417500617-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.762870 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.830283 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d65675a-7f0e-44b3-93df-5aad91061871" path="/var/lib/kubelet/pods/2d65675a-7f0e-44b3-93df-5aad91061871/volumes" Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.984939 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:13 crc kubenswrapper[4972]: I0228 10:58:13.995727 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.014862 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:14 crc kubenswrapper[4972]: E0228 10:58:14.015568 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="rabbitmq" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.015646 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="rabbitmq" Feb 28 10:58:14 crc kubenswrapper[4972]: E0228 10:58:14.015740 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="setup-container" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.015805 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="setup-container" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.016114 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" containerName="rabbitmq" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.017377 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.019715 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.020146 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zgbf4" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.020243 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.020360 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.020829 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.021074 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.021329 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.031074 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068087 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068168 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068247 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068288 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66540e7-d517-4b5d-bd92-efdd423596fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068313 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068352 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068408 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068479 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068512 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6jqf\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-kube-api-access-h6jqf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068545 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66540e7-d517-4b5d-bd92-efdd423596fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.068600 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.169899 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.169961 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6jqf\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-kube-api-access-h6jqf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.169989 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66540e7-d517-4b5d-bd92-efdd423596fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170029 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170055 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170072 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170098 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170154 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170184 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66540e7-d517-4b5d-bd92-efdd423596fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170205 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170224 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170264 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.170803 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.171237 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.172326 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.172352 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.172767 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66540e7-d517-4b5d-bd92-efdd423596fb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.175481 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66540e7-d517-4b5d-bd92-efdd423596fb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.175577 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.176225 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.176649 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66540e7-d517-4b5d-bd92-efdd423596fb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.199447 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6jqf\" (UniqueName: \"kubernetes.io/projected/b66540e7-d517-4b5d-bd92-efdd423596fb-kube-api-access-h6jqf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.233491 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66540e7-d517-4b5d-bd92-efdd423596fb\") " pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.340370 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.609360 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a821361-b8f7-4511-9f07-9f05f1b8425d","Type":"ContainerStarted","Data":"d7dfc5fae84853e9aef1858d0ea3e903843bc1bda2bb9c57494ea6cc0e67a549"} Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.618720 4972 generic.go:334] "Generic (PLEG): container finished" podID="886a175a-2921-4867-9d20-ebc327478bae" containerID="8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793" exitCode=0 Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.618778 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" event={"ID":"886a175a-2921-4867-9d20-ebc327478bae","Type":"ContainerDied","Data":"8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793"} Feb 28 10:58:14 crc kubenswrapper[4972]: W0228 10:58:14.822681 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb66540e7_d517_4b5d_bd92_efdd423596fb.slice/crio-637ab3d9751377e30eb150edd801252940aa406f2e24ec5885ad86dde6aae83c WatchSource:0}: Error finding container 637ab3d9751377e30eb150edd801252940aa406f2e24ec5885ad86dde6aae83c: Status 404 returned error can't find the container with id 637ab3d9751377e30eb150edd801252940aa406f2e24ec5885ad86dde6aae83c Feb 28 10:58:14 crc kubenswrapper[4972]: I0228 10:58:14.823290 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.629699 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" event={"ID":"886a175a-2921-4867-9d20-ebc327478bae","Type":"ContainerStarted","Data":"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02"} Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.630321 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.631825 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a821361-b8f7-4511-9f07-9f05f1b8425d","Type":"ContainerStarted","Data":"fb985ac54d4181c7f3b45eb2a9b425c051b81363cf11069bac7931bf8ccf547b"} Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.633279 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66540e7-d517-4b5d-bd92-efdd423596fb","Type":"ContainerStarted","Data":"637ab3d9751377e30eb150edd801252940aa406f2e24ec5885ad86dde6aae83c"} Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.668295 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" podStartSLOduration=3.668260762 podStartE2EDuration="3.668260762s" podCreationTimestamp="2026-02-28 10:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:58:15.652909804 +0000 UTC m=+1352.564917562" watchObservedRunningTime="2026-02-28 10:58:15.668260762 +0000 UTC m=+1352.580268520" Feb 28 10:58:15 crc kubenswrapper[4972]: I0228 10:58:15.806037 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffd40994-f4d6-442c-97fa-6ed417500617" path="/var/lib/kubelet/pods/ffd40994-f4d6-442c-97fa-6ed417500617/volumes" Feb 28 10:58:17 crc kubenswrapper[4972]: I0228 10:58:17.659977 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66540e7-d517-4b5d-bd92-efdd423596fb","Type":"ContainerStarted","Data":"7822ba1feff649fdcc13ebc63ae66cc3e9a60505cd275efe3951833f8164e4aa"} Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.046764 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.146215 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.146580 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="dnsmasq-dns" containerID="cri-o://5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401" gracePeriod=10 Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.411356 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-2qbql"] Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.413405 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.430187 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-2qbql"] Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488025 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-svc\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488089 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488300 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb8qt\" (UniqueName: \"kubernetes.io/projected/977a02d6-198e-4d62-a2aa-9f83ae367dc6-kube-api-access-mb8qt\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488333 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-config\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488673 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488719 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.488937 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596545 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596609 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596678 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596702 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-svc\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596726 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596789 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb8qt\" (UniqueName: \"kubernetes.io/projected/977a02d6-198e-4d62-a2aa-9f83ae367dc6-kube-api-access-mb8qt\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.596816 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-config\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.597883 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-config\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.598684 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-svc\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.599029 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.599405 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.599740 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.600382 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977a02d6-198e-4d62-a2aa-9f83ae367dc6-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.620367 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb8qt\" (UniqueName: \"kubernetes.io/projected/977a02d6-198e-4d62-a2aa-9f83ae367dc6-kube-api-access-mb8qt\") pod \"dnsmasq-dns-55478c4467-2qbql\" (UID: \"977a02d6-198e-4d62-a2aa-9f83ae367dc6\") " pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.715515 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.748159 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.759422 4972 generic.go:334] "Generic (PLEG): container finished" podID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerID="5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401" exitCode=0 Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.759492 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" event={"ID":"7fbf7692-27d2-4415-bd1d-ee7aace8f117","Type":"ContainerDied","Data":"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401"} Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.759526 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" event={"ID":"7fbf7692-27d2-4415-bd1d-ee7aace8f117","Type":"ContainerDied","Data":"dec37811ed815d5910673588b49c5056de5f50f2e64e56ada95e81cca652e343"} Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.759547 4972 scope.go:117] "RemoveContainer" containerID="5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.759562 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fk8xl" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.801816 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.802015 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs87g\" (UniqueName: \"kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.802091 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.802145 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.802204 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.802239 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0\") pod \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\" (UID: \"7fbf7692-27d2-4415-bd1d-ee7aace8f117\") " Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.811063 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g" (OuterVolumeSpecName: "kube-api-access-bs87g") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "kube-api-access-bs87g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.816034 4972 scope.go:117] "RemoveContainer" containerID="d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.854611 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.856129 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.866711 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config" (OuterVolumeSpecName: "config") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.870415 4972 scope.go:117] "RemoveContainer" containerID="5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401" Feb 28 10:58:23 crc kubenswrapper[4972]: E0228 10:58:23.871063 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401\": container with ID starting with 5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401 not found: ID does not exist" containerID="5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.871110 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401"} err="failed to get container status \"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401\": rpc error: code = NotFound desc = could not find container \"5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401\": container with ID starting with 5be47355af67203a97ea64a4885e467c814a8bc2575876edaa5eedfd3ddef401 not found: ID does not exist" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.871139 4972 scope.go:117] "RemoveContainer" containerID="d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267" Feb 28 10:58:23 crc kubenswrapper[4972]: E0228 10:58:23.871467 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267\": container with ID starting with d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267 not found: ID does not exist" containerID="d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.871502 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267"} err="failed to get container status \"d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267\": rpc error: code = NotFound desc = could not find container \"d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267\": container with ID starting with d098d0f1c04e3b101f8879f9afcdc34d9f8595f24f6ca9e5611e79a2e8f72267 not found: ID does not exist" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.883081 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.887266 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7fbf7692-27d2-4415-bd1d-ee7aace8f117" (UID: "7fbf7692-27d2-4415-bd1d-ee7aace8f117"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905394 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905423 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905433 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905442 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905452 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs87g\" (UniqueName: \"kubernetes.io/projected/7fbf7692-27d2-4415-bd1d-ee7aace8f117-kube-api-access-bs87g\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:23 crc kubenswrapper[4972]: I0228 10:58:23.905463 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fbf7692-27d2-4415-bd1d-ee7aace8f117-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.106380 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.119236 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fk8xl"] Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.266536 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-2qbql"] Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.780872 4972 generic.go:334] "Generic (PLEG): container finished" podID="977a02d6-198e-4d62-a2aa-9f83ae367dc6" containerID="87aa3f3e38653668254cc33d9ac90ab8fc8789e544d39232b59af72d416b9aff" exitCode=0 Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.781085 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-2qbql" event={"ID":"977a02d6-198e-4d62-a2aa-9f83ae367dc6","Type":"ContainerDied","Data":"87aa3f3e38653668254cc33d9ac90ab8fc8789e544d39232b59af72d416b9aff"} Feb 28 10:58:24 crc kubenswrapper[4972]: I0228 10:58:24.781356 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-2qbql" event={"ID":"977a02d6-198e-4d62-a2aa-9f83ae367dc6","Type":"ContainerStarted","Data":"bc475426c705ff8d96a7b19105ad923563b35112f065882920e4fd39cbc76499"} Feb 28 10:58:25 crc kubenswrapper[4972]: I0228 10:58:25.813123 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" path="/var/lib/kubelet/pods/7fbf7692-27d2-4415-bd1d-ee7aace8f117/volumes" Feb 28 10:58:25 crc kubenswrapper[4972]: I0228 10:58:25.815017 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:25 crc kubenswrapper[4972]: I0228 10:58:25.815100 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-2qbql" event={"ID":"977a02d6-198e-4d62-a2aa-9f83ae367dc6","Type":"ContainerStarted","Data":"3ce37981a4389e0f1a756fbf96bf50684dfaf810e48480f1deb6c3f5013fe351"} Feb 28 10:58:25 crc kubenswrapper[4972]: I0228 10:58:25.847308 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-2qbql" podStartSLOduration=2.847274805 podStartE2EDuration="2.847274805s" podCreationTimestamp="2026-02-28 10:58:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:58:25.839101217 +0000 UTC m=+1362.751108995" watchObservedRunningTime="2026-02-28 10:58:25.847274805 +0000 UTC m=+1362.759282583" Feb 28 10:58:28 crc kubenswrapper[4972]: I0228 10:58:28.890597 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 10:58:28 crc kubenswrapper[4972]: I0228 10:58:28.891682 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 10:58:28 crc kubenswrapper[4972]: I0228 10:58:28.891770 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 10:58:28 crc kubenswrapper[4972]: I0228 10:58:28.893011 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 10:58:28 crc kubenswrapper[4972]: I0228 10:58:28.893108 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8" gracePeriod=600 Feb 28 10:58:29 crc kubenswrapper[4972]: I0228 10:58:29.854019 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8" exitCode=0 Feb 28 10:58:29 crc kubenswrapper[4972]: I0228 10:58:29.854133 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8"} Feb 28 10:58:29 crc kubenswrapper[4972]: I0228 10:58:29.854774 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0"} Feb 28 10:58:29 crc kubenswrapper[4972]: I0228 10:58:29.854811 4972 scope.go:117] "RemoveContainer" containerID="9fe43270f2d63e70af42e0ed5ff252c29e540ced678fee8f8cbb8bdc728a0e80" Feb 28 10:58:33 crc kubenswrapper[4972]: I0228 10:58:33.754907 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-2qbql" Feb 28 10:58:33 crc kubenswrapper[4972]: I0228 10:58:33.862879 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:33 crc kubenswrapper[4972]: I0228 10:58:33.863701 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="dnsmasq-dns" containerID="cri-o://aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02" gracePeriod=10 Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.318647 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468173 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468236 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468262 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468593 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468619 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468672 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlq94\" (UniqueName: \"kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.468702 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc\") pod \"886a175a-2921-4867-9d20-ebc327478bae\" (UID: \"886a175a-2921-4867-9d20-ebc327478bae\") " Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.488764 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94" (OuterVolumeSpecName: "kube-api-access-wlq94") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "kube-api-access-wlq94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.521331 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.522379 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.528122 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.532784 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config" (OuterVolumeSpecName: "config") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.539187 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.548111 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "886a175a-2921-4867-9d20-ebc327478bae" (UID: "886a175a-2921-4867-9d20-ebc327478bae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573228 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573267 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573280 4972 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573290 4972 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-config\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573299 4972 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573307 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlq94\" (UniqueName: \"kubernetes.io/projected/886a175a-2921-4867-9d20-ebc327478bae-kube-api-access-wlq94\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.573318 4972 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/886a175a-2921-4867-9d20-ebc327478bae-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.918368 4972 generic.go:334] "Generic (PLEG): container finished" podID="886a175a-2921-4867-9d20-ebc327478bae" containerID="aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02" exitCode=0 Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.918498 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.918534 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" event={"ID":"886a175a-2921-4867-9d20-ebc327478bae","Type":"ContainerDied","Data":"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02"} Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.919173 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-hhpsd" event={"ID":"886a175a-2921-4867-9d20-ebc327478bae","Type":"ContainerDied","Data":"073a8dd505de434f55bdcdcbdd0292512d67a61ec22179ee5b383c50de032a6a"} Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.919218 4972 scope.go:117] "RemoveContainer" containerID="aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.947653 4972 scope.go:117] "RemoveContainer" containerID="8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.971088 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.979626 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-hhpsd"] Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.982777 4972 scope.go:117] "RemoveContainer" containerID="aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02" Feb 28 10:58:34 crc kubenswrapper[4972]: E0228 10:58:34.983270 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02\": container with ID starting with aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02 not found: ID does not exist" containerID="aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.983317 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02"} err="failed to get container status \"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02\": rpc error: code = NotFound desc = could not find container \"aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02\": container with ID starting with aa0fbc3a3a7176fae512355027db8446be92a66a907156d717ffe3c981b33a02 not found: ID does not exist" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.983349 4972 scope.go:117] "RemoveContainer" containerID="8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793" Feb 28 10:58:34 crc kubenswrapper[4972]: E0228 10:58:34.983913 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793\": container with ID starting with 8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793 not found: ID does not exist" containerID="8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793" Feb 28 10:58:34 crc kubenswrapper[4972]: I0228 10:58:34.983945 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793"} err="failed to get container status \"8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793\": rpc error: code = NotFound desc = could not find container \"8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793\": container with ID starting with 8f58d7ba4bb4e10dd1e9c7adbe10dbaa70e4bd98d17c0d1fc40325b03413a793 not found: ID does not exist" Feb 28 10:58:35 crc kubenswrapper[4972]: I0228 10:58:35.806479 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="886a175a-2921-4867-9d20-ebc327478bae" path="/var/lib/kubelet/pods/886a175a-2921-4867-9d20-ebc327478bae/volumes" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.125662 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f"] Feb 28 10:58:46 crc kubenswrapper[4972]: E0228 10:58:46.126686 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126700 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: E0228 10:58:46.126711 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="init" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126717 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="init" Feb 28 10:58:46 crc kubenswrapper[4972]: E0228 10:58:46.126731 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126738 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: E0228 10:58:46.126752 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="init" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126757 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="init" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126933 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbf7692-27d2-4415-bd1d-ee7aace8f117" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.126946 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="886a175a-2921-4867-9d20-ebc327478bae" containerName="dnsmasq-dns" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.127652 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.131005 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.131301 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.131353 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.131540 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.139191 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f"] Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.245330 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.245389 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.245424 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.245562 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4r9t\" (UniqueName: \"kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.347577 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.347675 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.347741 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.347818 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4r9t\" (UniqueName: \"kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.356743 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.358767 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.360134 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.367200 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4r9t\" (UniqueName: \"kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:46 crc kubenswrapper[4972]: I0228 10:58:46.463979 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:58:47 crc kubenswrapper[4972]: I0228 10:58:47.147916 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f"] Feb 28 10:58:48 crc kubenswrapper[4972]: I0228 10:58:48.084703 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" event={"ID":"a87a4176-9953-45b9-8db1-824b39dd2e4d","Type":"ContainerStarted","Data":"88cafe4f0d16d2e0cde7c7e61a8819e9c08c571aa59cbc7a13f556b611738fd7"} Feb 28 10:58:48 crc kubenswrapper[4972]: E0228 10:58:48.306992 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a821361_b8f7_4511_9f07_9f05f1b8425d.slice/crio-fb985ac54d4181c7f3b45eb2a9b425c051b81363cf11069bac7931bf8ccf547b.scope\": RecentStats: unable to find data in memory cache]" Feb 28 10:58:49 crc kubenswrapper[4972]: I0228 10:58:49.101677 4972 generic.go:334] "Generic (PLEG): container finished" podID="6a821361-b8f7-4511-9f07-9f05f1b8425d" containerID="fb985ac54d4181c7f3b45eb2a9b425c051b81363cf11069bac7931bf8ccf547b" exitCode=0 Feb 28 10:58:49 crc kubenswrapper[4972]: I0228 10:58:49.101826 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a821361-b8f7-4511-9f07-9f05f1b8425d","Type":"ContainerDied","Data":"fb985ac54d4181c7f3b45eb2a9b425c051b81363cf11069bac7931bf8ccf547b"} Feb 28 10:58:50 crc kubenswrapper[4972]: I0228 10:58:50.114281 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a821361-b8f7-4511-9f07-9f05f1b8425d","Type":"ContainerStarted","Data":"857ba541f028f7d042dfb1e711cff2086f1dee46cf9c2fe156198d80795bfedb"} Feb 28 10:58:50 crc kubenswrapper[4972]: I0228 10:58:50.114792 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 28 10:58:50 crc kubenswrapper[4972]: I0228 10:58:50.120133 4972 generic.go:334] "Generic (PLEG): container finished" podID="b66540e7-d517-4b5d-bd92-efdd423596fb" containerID="7822ba1feff649fdcc13ebc63ae66cc3e9a60505cd275efe3951833f8164e4aa" exitCode=0 Feb 28 10:58:50 crc kubenswrapper[4972]: I0228 10:58:50.120195 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66540e7-d517-4b5d-bd92-efdd423596fb","Type":"ContainerDied","Data":"7822ba1feff649fdcc13ebc63ae66cc3e9a60505cd275efe3951833f8164e4aa"} Feb 28 10:58:50 crc kubenswrapper[4972]: I0228 10:58:50.156592 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.156563361 podStartE2EDuration="38.156563361s" podCreationTimestamp="2026-02-28 10:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:58:50.14610132 +0000 UTC m=+1387.058109058" watchObservedRunningTime="2026-02-28 10:58:50.156563361 +0000 UTC m=+1387.068571109" Feb 28 10:58:51 crc kubenswrapper[4972]: I0228 10:58:51.136256 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66540e7-d517-4b5d-bd92-efdd423596fb","Type":"ContainerStarted","Data":"708c8b528281d4edc7cbeb4e1cffb98fa90bbc75d97f4e9f9e7bb5c04d9d6e39"} Feb 28 10:58:51 crc kubenswrapper[4972]: I0228 10:58:51.137254 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:58:51 crc kubenswrapper[4972]: I0228 10:58:51.170866 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.170843348 podStartE2EDuration="38.170843348s" podCreationTimestamp="2026-02-28 10:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 10:58:51.156053145 +0000 UTC m=+1388.068060893" watchObservedRunningTime="2026-02-28 10:58:51.170843348 +0000 UTC m=+1388.082851086" Feb 28 10:58:59 crc kubenswrapper[4972]: I0228 10:58:59.234350 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" event={"ID":"a87a4176-9953-45b9-8db1-824b39dd2e4d","Type":"ContainerStarted","Data":"fa94d9eab32e2cfbedef46e6619efbb7dfd06638720a512ff77ebb6070ee12e6"} Feb 28 10:58:59 crc kubenswrapper[4972]: I0228 10:58:59.274190 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" podStartSLOduration=2.020901629 podStartE2EDuration="13.274120841s" podCreationTimestamp="2026-02-28 10:58:46 +0000 UTC" firstStartedPulling="2026-02-28 10:58:47.153276128 +0000 UTC m=+1384.065283866" lastFinishedPulling="2026-02-28 10:58:58.40649534 +0000 UTC m=+1395.318503078" observedRunningTime="2026-02-28 10:58:59.262828515 +0000 UTC m=+1396.174836253" watchObservedRunningTime="2026-02-28 10:58:59.274120841 +0000 UTC m=+1396.186128589" Feb 28 10:58:59 crc kubenswrapper[4972]: I0228 10:58:59.735019 4972 scope.go:117] "RemoveContainer" containerID="e03ec78b341831c5f5f29feb429aef7b566d4988575ed872c29341e63e13117e" Feb 28 10:59:03 crc kubenswrapper[4972]: I0228 10:59:03.263703 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 28 10:59:04 crc kubenswrapper[4972]: I0228 10:59:04.344974 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 28 10:59:11 crc kubenswrapper[4972]: I0228 10:59:11.384817 4972 generic.go:334] "Generic (PLEG): container finished" podID="a87a4176-9953-45b9-8db1-824b39dd2e4d" containerID="fa94d9eab32e2cfbedef46e6619efbb7dfd06638720a512ff77ebb6070ee12e6" exitCode=0 Feb 28 10:59:11 crc kubenswrapper[4972]: I0228 10:59:11.384889 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" event={"ID":"a87a4176-9953-45b9-8db1-824b39dd2e4d","Type":"ContainerDied","Data":"fa94d9eab32e2cfbedef46e6619efbb7dfd06638720a512ff77ebb6070ee12e6"} Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.140747 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.142577 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.166663 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.236386 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.236646 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqpnd\" (UniqueName: \"kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.236742 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.338610 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.338756 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.338856 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqpnd\" (UniqueName: \"kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.339199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.339263 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.358510 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqpnd\" (UniqueName: \"kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd\") pod \"redhat-operators-pqcvr\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.462107 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:12 crc kubenswrapper[4972]: I0228 10:59:12.940809 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.044215 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.054296 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle\") pod \"a87a4176-9953-45b9-8db1-824b39dd2e4d\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.054559 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4r9t\" (UniqueName: \"kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t\") pod \"a87a4176-9953-45b9-8db1-824b39dd2e4d\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.054673 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory\") pod \"a87a4176-9953-45b9-8db1-824b39dd2e4d\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.054769 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam\") pod \"a87a4176-9953-45b9-8db1-824b39dd2e4d\" (UID: \"a87a4176-9953-45b9-8db1-824b39dd2e4d\") " Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.062353 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t" (OuterVolumeSpecName: "kube-api-access-r4r9t") pod "a87a4176-9953-45b9-8db1-824b39dd2e4d" (UID: "a87a4176-9953-45b9-8db1-824b39dd2e4d"). InnerVolumeSpecName "kube-api-access-r4r9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.065124 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a87a4176-9953-45b9-8db1-824b39dd2e4d" (UID: "a87a4176-9953-45b9-8db1-824b39dd2e4d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.102041 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a87a4176-9953-45b9-8db1-824b39dd2e4d" (UID: "a87a4176-9953-45b9-8db1-824b39dd2e4d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.106612 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory" (OuterVolumeSpecName: "inventory") pod "a87a4176-9953-45b9-8db1-824b39dd2e4d" (UID: "a87a4176-9953-45b9-8db1-824b39dd2e4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.157267 4972 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.157311 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4r9t\" (UniqueName: \"kubernetes.io/projected/a87a4176-9953-45b9-8db1-824b39dd2e4d-kube-api-access-r4r9t\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.157322 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.157332 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a87a4176-9953-45b9-8db1-824b39dd2e4d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.409751 4972 generic.go:334] "Generic (PLEG): container finished" podID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerID="094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326" exitCode=0 Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.409820 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerDied","Data":"094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326"} Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.409848 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerStarted","Data":"3b97dc769b9821cb7c7740786e04059583ee18831fb12b67289288effc6c47ae"} Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.412199 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.420309 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" event={"ID":"a87a4176-9953-45b9-8db1-824b39dd2e4d","Type":"ContainerDied","Data":"88cafe4f0d16d2e0cde7c7e61a8819e9c08c571aa59cbc7a13f556b611738fd7"} Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.420379 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88cafe4f0d16d2e0cde7c7e61a8819e9c08c571aa59cbc7a13f556b611738fd7" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.420623 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.563110 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2"] Feb 28 10:59:13 crc kubenswrapper[4972]: E0228 10:59:13.564026 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87a4176-9953-45b9-8db1-824b39dd2e4d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.564045 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87a4176-9953-45b9-8db1-824b39dd2e4d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.564240 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87a4176-9953-45b9-8db1-824b39dd2e4d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.565066 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.569513 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.569567 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.569592 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.570991 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.611639 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2"] Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.666511 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.666597 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.666659 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpxzh\" (UniqueName: \"kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.769846 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.769969 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.770072 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpxzh\" (UniqueName: \"kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.780378 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.784228 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.801791 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpxzh\" (UniqueName: \"kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bvts2\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:13 crc kubenswrapper[4972]: I0228 10:59:13.896252 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:14 crc kubenswrapper[4972]: I0228 10:59:14.433865 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerStarted","Data":"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d"} Feb 28 10:59:14 crc kubenswrapper[4972]: W0228 10:59:14.488223 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6c60946_e11b_4644_8742_fb1225f8d690.slice/crio-53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b WatchSource:0}: Error finding container 53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b: Status 404 returned error can't find the container with id 53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b Feb 28 10:59:14 crc kubenswrapper[4972]: I0228 10:59:14.499517 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2"] Feb 28 10:59:15 crc kubenswrapper[4972]: I0228 10:59:15.442991 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" event={"ID":"a6c60946-e11b-4644-8742-fb1225f8d690","Type":"ContainerStarted","Data":"53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b"} Feb 28 10:59:16 crc kubenswrapper[4972]: I0228 10:59:16.457252 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" event={"ID":"a6c60946-e11b-4644-8742-fb1225f8d690","Type":"ContainerStarted","Data":"8d2ff41269fbc8e369fd3761e1e2fcdc74ce056e43960496bd85a83e1904ab54"} Feb 28 10:59:16 crc kubenswrapper[4972]: I0228 10:59:16.486874 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" podStartSLOduration=2.706974838 podStartE2EDuration="3.486849877s" podCreationTimestamp="2026-02-28 10:59:13 +0000 UTC" firstStartedPulling="2026-02-28 10:59:14.493042606 +0000 UTC m=+1411.405050344" lastFinishedPulling="2026-02-28 10:59:15.272917605 +0000 UTC m=+1412.184925383" observedRunningTime="2026-02-28 10:59:16.485074318 +0000 UTC m=+1413.397082056" watchObservedRunningTime="2026-02-28 10:59:16.486849877 +0000 UTC m=+1413.398857615" Feb 28 10:59:17 crc kubenswrapper[4972]: I0228 10:59:17.470629 4972 generic.go:334] "Generic (PLEG): container finished" podID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerID="eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d" exitCode=0 Feb 28 10:59:17 crc kubenswrapper[4972]: I0228 10:59:17.470757 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerDied","Data":"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d"} Feb 28 10:59:18 crc kubenswrapper[4972]: I0228 10:59:18.487735 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerStarted","Data":"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff"} Feb 28 10:59:18 crc kubenswrapper[4972]: I0228 10:59:18.520703 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pqcvr" podStartSLOduration=2.071859222 podStartE2EDuration="6.520674346s" podCreationTimestamp="2026-02-28 10:59:12 +0000 UTC" firstStartedPulling="2026-02-28 10:59:13.412010855 +0000 UTC m=+1410.324018593" lastFinishedPulling="2026-02-28 10:59:17.860825979 +0000 UTC m=+1414.772833717" observedRunningTime="2026-02-28 10:59:18.514594176 +0000 UTC m=+1415.426601924" watchObservedRunningTime="2026-02-28 10:59:18.520674346 +0000 UTC m=+1415.432682084" Feb 28 10:59:19 crc kubenswrapper[4972]: I0228 10:59:19.515267 4972 generic.go:334] "Generic (PLEG): container finished" podID="a6c60946-e11b-4644-8742-fb1225f8d690" containerID="8d2ff41269fbc8e369fd3761e1e2fcdc74ce056e43960496bd85a83e1904ab54" exitCode=0 Feb 28 10:59:19 crc kubenswrapper[4972]: I0228 10:59:19.515371 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" event={"ID":"a6c60946-e11b-4644-8742-fb1225f8d690","Type":"ContainerDied","Data":"8d2ff41269fbc8e369fd3761e1e2fcdc74ce056e43960496bd85a83e1904ab54"} Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.067111 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.157316 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam\") pod \"a6c60946-e11b-4644-8742-fb1225f8d690\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.157443 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpxzh\" (UniqueName: \"kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh\") pod \"a6c60946-e11b-4644-8742-fb1225f8d690\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.157678 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory\") pod \"a6c60946-e11b-4644-8742-fb1225f8d690\" (UID: \"a6c60946-e11b-4644-8742-fb1225f8d690\") " Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.183758 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh" (OuterVolumeSpecName: "kube-api-access-rpxzh") pod "a6c60946-e11b-4644-8742-fb1225f8d690" (UID: "a6c60946-e11b-4644-8742-fb1225f8d690"). InnerVolumeSpecName "kube-api-access-rpxzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.205583 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory" (OuterVolumeSpecName: "inventory") pod "a6c60946-e11b-4644-8742-fb1225f8d690" (UID: "a6c60946-e11b-4644-8742-fb1225f8d690"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.206875 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a6c60946-e11b-4644-8742-fb1225f8d690" (UID: "a6c60946-e11b-4644-8742-fb1225f8d690"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.275275 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.275621 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpxzh\" (UniqueName: \"kubernetes.io/projected/a6c60946-e11b-4644-8742-fb1225f8d690-kube-api-access-rpxzh\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.275635 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6c60946-e11b-4644-8742-fb1225f8d690-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.540281 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" event={"ID":"a6c60946-e11b-4644-8742-fb1225f8d690","Type":"ContainerDied","Data":"53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b"} Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.540358 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53322393c66f3becfc54951cb88afd0de24c0140365f9021623c78620fa4250b" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.540512 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bvts2" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.752055 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc"] Feb 28 10:59:21 crc kubenswrapper[4972]: E0228 10:59:21.752671 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6c60946-e11b-4644-8742-fb1225f8d690" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.752695 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6c60946-e11b-4644-8742-fb1225f8d690" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.762571 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6c60946-e11b-4644-8742-fb1225f8d690" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.763716 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.770141 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.770348 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.770502 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.770684 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.780050 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc"] Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.888076 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.888151 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzdbp\" (UniqueName: \"kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.888258 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.888286 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.989520 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.989629 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzdbp\" (UniqueName: \"kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.989787 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.989827 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.996975 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.996990 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:21 crc kubenswrapper[4972]: I0228 10:59:21.998521 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:22 crc kubenswrapper[4972]: I0228 10:59:22.024689 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzdbp\" (UniqueName: \"kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:22 crc kubenswrapper[4972]: I0228 10:59:22.087349 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 10:59:22 crc kubenswrapper[4972]: I0228 10:59:22.462686 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:22 crc kubenswrapper[4972]: I0228 10:59:22.463155 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:22 crc kubenswrapper[4972]: I0228 10:59:22.691553 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc"] Feb 28 10:59:22 crc kubenswrapper[4972]: W0228 10:59:22.693762 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9d8a507_2de3_4624_ad38_5d50c0c50874.slice/crio-947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9 WatchSource:0}: Error finding container 947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9: Status 404 returned error can't find the container with id 947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9 Feb 28 10:59:23 crc kubenswrapper[4972]: I0228 10:59:23.522819 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pqcvr" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="registry-server" probeResult="failure" output=< Feb 28 10:59:23 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 10:59:23 crc kubenswrapper[4972]: > Feb 28 10:59:23 crc kubenswrapper[4972]: I0228 10:59:23.562977 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" event={"ID":"e9d8a507-2de3-4624-ad38-5d50c0c50874","Type":"ContainerStarted","Data":"342182c25f2ecb20b83bda8c3bfd6632745a41414febfb3c7f083fa12be90e68"} Feb 28 10:59:23 crc kubenswrapper[4972]: I0228 10:59:23.563042 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" event={"ID":"e9d8a507-2de3-4624-ad38-5d50c0c50874","Type":"ContainerStarted","Data":"947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9"} Feb 28 10:59:23 crc kubenswrapper[4972]: I0228 10:59:23.582986 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" podStartSLOduration=2.157173902 podStartE2EDuration="2.582965293s" podCreationTimestamp="2026-02-28 10:59:21 +0000 UTC" firstStartedPulling="2026-02-28 10:59:22.697777092 +0000 UTC m=+1419.609784830" lastFinishedPulling="2026-02-28 10:59:23.123568483 +0000 UTC m=+1420.035576221" observedRunningTime="2026-02-28 10:59:23.582569451 +0000 UTC m=+1420.494577189" watchObservedRunningTime="2026-02-28 10:59:23.582965293 +0000 UTC m=+1420.494973021" Feb 28 10:59:32 crc kubenswrapper[4972]: I0228 10:59:32.530485 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:32 crc kubenswrapper[4972]: I0228 10:59:32.595422 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:32 crc kubenswrapper[4972]: I0228 10:59:32.772554 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:33 crc kubenswrapper[4972]: I0228 10:59:33.700630 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pqcvr" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="registry-server" containerID="cri-o://6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff" gracePeriod=2 Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.244494 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.416392 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities\") pod \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.416509 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqpnd\" (UniqueName: \"kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd\") pod \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.416581 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content\") pod \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\" (UID: \"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6\") " Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.417417 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities" (OuterVolumeSpecName: "utilities") pod "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" (UID: "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.423386 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd" (OuterVolumeSpecName: "kube-api-access-wqpnd") pod "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" (UID: "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6"). InnerVolumeSpecName "kube-api-access-wqpnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.519513 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.519990 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqpnd\" (UniqueName: \"kubernetes.io/projected/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-kube-api-access-wqpnd\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.536149 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" (UID: "de02ac5a-c33e-4ef1-9f49-e0b730ac58f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.622744 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.715876 4972 generic.go:334] "Generic (PLEG): container finished" podID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerID="6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff" exitCode=0 Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.715941 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqcvr" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.715945 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerDied","Data":"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff"} Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.716377 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqcvr" event={"ID":"de02ac5a-c33e-4ef1-9f49-e0b730ac58f6","Type":"ContainerDied","Data":"3b97dc769b9821cb7c7740786e04059583ee18831fb12b67289288effc6c47ae"} Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.716453 4972 scope.go:117] "RemoveContainer" containerID="6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.760102 4972 scope.go:117] "RemoveContainer" containerID="eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.771613 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.783441 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pqcvr"] Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.795956 4972 scope.go:117] "RemoveContainer" containerID="094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.846049 4972 scope.go:117] "RemoveContainer" containerID="6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff" Feb 28 10:59:34 crc kubenswrapper[4972]: E0228 10:59:34.846689 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff\": container with ID starting with 6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff not found: ID does not exist" containerID="6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.846730 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff"} err="failed to get container status \"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff\": rpc error: code = NotFound desc = could not find container \"6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff\": container with ID starting with 6a0182e9ee05355e50b60fbc0b7c272ba242314c4fc853323f05789a97c5fbff not found: ID does not exist" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.846760 4972 scope.go:117] "RemoveContainer" containerID="eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d" Feb 28 10:59:34 crc kubenswrapper[4972]: E0228 10:59:34.847631 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d\": container with ID starting with eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d not found: ID does not exist" containerID="eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.847662 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d"} err="failed to get container status \"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d\": rpc error: code = NotFound desc = could not find container \"eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d\": container with ID starting with eaa9e5a8fb353162fe172a40984c46ff77b938065d7178398b090abd7a2df51d not found: ID does not exist" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.847677 4972 scope.go:117] "RemoveContainer" containerID="094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326" Feb 28 10:59:34 crc kubenswrapper[4972]: E0228 10:59:34.848292 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326\": container with ID starting with 094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326 not found: ID does not exist" containerID="094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326" Feb 28 10:59:34 crc kubenswrapper[4972]: I0228 10:59:34.848373 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326"} err="failed to get container status \"094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326\": rpc error: code = NotFound desc = could not find container \"094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326\": container with ID starting with 094093353b51581761d2e3dd1ebef15e4bffc53f247e422a8f3221c4af3b7326 not found: ID does not exist" Feb 28 10:59:35 crc kubenswrapper[4972]: I0228 10:59:35.803861 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" path="/var/lib/kubelet/pods/de02ac5a-c33e-4ef1-9f49-e0b730ac58f6/volumes" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.003865 4972 scope.go:117] "RemoveContainer" containerID="d14170a389bd3189c90e14bac64852e4ece03ac8764a700c6394d2fd33c6569d" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.037161 4972 scope.go:117] "RemoveContainer" containerID="cfa79cc5821a935aab1b950283231368dc1cdc570e9ccc09d78463e7e2f02f96" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.089747 4972 scope.go:117] "RemoveContainer" containerID="23d811b848cda59561d05b9cd5ee90bd85b638ded4a8d0b3e3841f44714313da" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.230581 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz"] Feb 28 11:00:00 crc kubenswrapper[4972]: E0228 11:00:00.231233 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="extract-content" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.231254 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="extract-content" Feb 28 11:00:00 crc kubenswrapper[4972]: E0228 11:00:00.231267 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="registry-server" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.231275 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="registry-server" Feb 28 11:00:00 crc kubenswrapper[4972]: E0228 11:00:00.231292 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="extract-utilities" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.231298 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="extract-utilities" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.231527 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="de02ac5a-c33e-4ef1-9f49-e0b730ac58f6" containerName="registry-server" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.232654 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.242925 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz"] Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.243531 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.243800 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.317205 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.317756 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.317860 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sxgr\" (UniqueName: \"kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.318764 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537940-mgnwl"] Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.326286 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.330161 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537940-mgnwl"] Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.338517 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.338797 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.351939 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.419544 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sxgr\" (UniqueName: \"kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.419651 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x88xl\" (UniqueName: \"kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl\") pod \"auto-csr-approver-29537940-mgnwl\" (UID: \"7766b68b-d801-4a78-b664-9ef12c2d30c0\") " pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.419685 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.419739 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.420755 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.439437 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.443183 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sxgr\" (UniqueName: \"kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr\") pod \"collect-profiles-29537940-6xkzz\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.521391 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x88xl\" (UniqueName: \"kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl\") pod \"auto-csr-approver-29537940-mgnwl\" (UID: \"7766b68b-d801-4a78-b664-9ef12c2d30c0\") " pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.537886 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x88xl\" (UniqueName: \"kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl\") pod \"auto-csr-approver-29537940-mgnwl\" (UID: \"7766b68b-d801-4a78-b664-9ef12c2d30c0\") " pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.572930 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:00 crc kubenswrapper[4972]: I0228 11:00:00.675866 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:01 crc kubenswrapper[4972]: I0228 11:00:01.055024 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz"] Feb 28 11:00:01 crc kubenswrapper[4972]: I0228 11:00:01.153415 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537940-mgnwl"] Feb 28 11:00:01 crc kubenswrapper[4972]: W0228 11:00:01.155889 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7766b68b_d801_4a78_b664_9ef12c2d30c0.slice/crio-c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d WatchSource:0}: Error finding container c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d: Status 404 returned error can't find the container with id c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d Feb 28 11:00:02 crc kubenswrapper[4972]: I0228 11:00:02.076482 4972 generic.go:334] "Generic (PLEG): container finished" podID="8edbd92b-6e5b-4c0f-b5d0-517a31b06960" containerID="ae89a2b6508bbc454200bbf4bfcdce2c7398ae70184ef633f01be25849c96dee" exitCode=0 Feb 28 11:00:02 crc kubenswrapper[4972]: I0228 11:00:02.077255 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" event={"ID":"8edbd92b-6e5b-4c0f-b5d0-517a31b06960","Type":"ContainerDied","Data":"ae89a2b6508bbc454200bbf4bfcdce2c7398ae70184ef633f01be25849c96dee"} Feb 28 11:00:02 crc kubenswrapper[4972]: I0228 11:00:02.077331 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" event={"ID":"8edbd92b-6e5b-4c0f-b5d0-517a31b06960","Type":"ContainerStarted","Data":"608b1aaf96eea207f04fb71bf3a0f5823c730aa09f88f74ff4c1e4fb1c60cfd2"} Feb 28 11:00:02 crc kubenswrapper[4972]: I0228 11:00:02.079471 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" event={"ID":"7766b68b-d801-4a78-b664-9ef12c2d30c0","Type":"ContainerStarted","Data":"c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d"} Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.457075 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.588183 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sxgr\" (UniqueName: \"kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr\") pod \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.588250 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume\") pod \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.588362 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume\") pod \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\" (UID: \"8edbd92b-6e5b-4c0f-b5d0-517a31b06960\") " Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.589622 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume" (OuterVolumeSpecName: "config-volume") pod "8edbd92b-6e5b-4c0f-b5d0-517a31b06960" (UID: "8edbd92b-6e5b-4c0f-b5d0-517a31b06960"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.596362 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8edbd92b-6e5b-4c0f-b5d0-517a31b06960" (UID: "8edbd92b-6e5b-4c0f-b5d0-517a31b06960"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.597222 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr" (OuterVolumeSpecName: "kube-api-access-4sxgr") pod "8edbd92b-6e5b-4c0f-b5d0-517a31b06960" (UID: "8edbd92b-6e5b-4c0f-b5d0-517a31b06960"). InnerVolumeSpecName "kube-api-access-4sxgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.690654 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.690693 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sxgr\" (UniqueName: \"kubernetes.io/projected/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-kube-api-access-4sxgr\") on node \"crc\" DevicePath \"\"" Feb 28 11:00:03 crc kubenswrapper[4972]: I0228 11:00:03.690708 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edbd92b-6e5b-4c0f-b5d0-517a31b06960-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:00:04 crc kubenswrapper[4972]: I0228 11:00:04.111004 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" Feb 28 11:00:04 crc kubenswrapper[4972]: I0228 11:00:04.110995 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz" event={"ID":"8edbd92b-6e5b-4c0f-b5d0-517a31b06960","Type":"ContainerDied","Data":"608b1aaf96eea207f04fb71bf3a0f5823c730aa09f88f74ff4c1e4fb1c60cfd2"} Feb 28 11:00:04 crc kubenswrapper[4972]: I0228 11:00:04.111140 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="608b1aaf96eea207f04fb71bf3a0f5823c730aa09f88f74ff4c1e4fb1c60cfd2" Feb 28 11:00:04 crc kubenswrapper[4972]: I0228 11:00:04.112972 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" event={"ID":"7766b68b-d801-4a78-b664-9ef12c2d30c0","Type":"ContainerStarted","Data":"af27305d025d2829b00a450c674e3957338dcee34e93cfec610ec7229627c73c"} Feb 28 11:00:04 crc kubenswrapper[4972]: I0228 11:00:04.136622 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" podStartSLOduration=1.637600824 podStartE2EDuration="4.136606952s" podCreationTimestamp="2026-02-28 11:00:00 +0000 UTC" firstStartedPulling="2026-02-28 11:00:01.159699048 +0000 UTC m=+1458.071706796" lastFinishedPulling="2026-02-28 11:00:03.658705186 +0000 UTC m=+1460.570712924" observedRunningTime="2026-02-28 11:00:04.126784142 +0000 UTC m=+1461.038791880" watchObservedRunningTime="2026-02-28 11:00:04.136606952 +0000 UTC m=+1461.048614690" Feb 28 11:00:05 crc kubenswrapper[4972]: I0228 11:00:05.127822 4972 generic.go:334] "Generic (PLEG): container finished" podID="7766b68b-d801-4a78-b664-9ef12c2d30c0" containerID="af27305d025d2829b00a450c674e3957338dcee34e93cfec610ec7229627c73c" exitCode=0 Feb 28 11:00:05 crc kubenswrapper[4972]: I0228 11:00:05.127877 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" event={"ID":"7766b68b-d801-4a78-b664-9ef12c2d30c0","Type":"ContainerDied","Data":"af27305d025d2829b00a450c674e3957338dcee34e93cfec610ec7229627c73c"} Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.534004 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.658943 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x88xl\" (UniqueName: \"kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl\") pod \"7766b68b-d801-4a78-b664-9ef12c2d30c0\" (UID: \"7766b68b-d801-4a78-b664-9ef12c2d30c0\") " Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.669865 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl" (OuterVolumeSpecName: "kube-api-access-x88xl") pod "7766b68b-d801-4a78-b664-9ef12c2d30c0" (UID: "7766b68b-d801-4a78-b664-9ef12c2d30c0"). InnerVolumeSpecName "kube-api-access-x88xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.761360 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x88xl\" (UniqueName: \"kubernetes.io/projected/7766b68b-d801-4a78-b664-9ef12c2d30c0-kube-api-access-x88xl\") on node \"crc\" DevicePath \"\"" Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.902665 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537934-2kjkh"] Feb 28 11:00:06 crc kubenswrapper[4972]: I0228 11:00:06.911522 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537934-2kjkh"] Feb 28 11:00:07 crc kubenswrapper[4972]: I0228 11:00:07.153379 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" event={"ID":"7766b68b-d801-4a78-b664-9ef12c2d30c0","Type":"ContainerDied","Data":"c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d"} Feb 28 11:00:07 crc kubenswrapper[4972]: I0228 11:00:07.153422 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9fbf344eef639295cac5488da0c4315a0b85ed05a6765292743ff80c6c7202d" Feb 28 11:00:07 crc kubenswrapper[4972]: I0228 11:00:07.153489 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537940-mgnwl" Feb 28 11:00:07 crc kubenswrapper[4972]: I0228 11:00:07.801143 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="625f85df-e9dd-4c46-9a05-b4da0c0bd5a3" path="/var/lib/kubelet/pods/625f85df-e9dd-4c46-9a05-b4da0c0bd5a3/volumes" Feb 28 11:00:58 crc kubenswrapper[4972]: I0228 11:00:58.891165 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:00:58 crc kubenswrapper[4972]: I0228 11:00:58.891916 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.187590 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29537941-8tjbq"] Feb 28 11:01:00 crc kubenswrapper[4972]: E0228 11:01:00.188363 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7766b68b-d801-4a78-b664-9ef12c2d30c0" containerName="oc" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.188377 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7766b68b-d801-4a78-b664-9ef12c2d30c0" containerName="oc" Feb 28 11:01:00 crc kubenswrapper[4972]: E0228 11:01:00.188387 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edbd92b-6e5b-4c0f-b5d0-517a31b06960" containerName="collect-profiles" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.188394 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edbd92b-6e5b-4c0f-b5d0-517a31b06960" containerName="collect-profiles" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.188596 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edbd92b-6e5b-4c0f-b5d0-517a31b06960" containerName="collect-profiles" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.188617 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="7766b68b-d801-4a78-b664-9ef12c2d30c0" containerName="oc" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.189354 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.200107 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29537941-8tjbq"] Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.338258 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.338594 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.338695 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2tx\" (UniqueName: \"kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.338841 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.435806 4972 scope.go:117] "RemoveContainer" containerID="8be33d659d45e2374c60706378eb0db0668124a28ee7b3a2e70f6638aa5acd42" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.442157 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.442241 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.442267 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2tx\" (UniqueName: \"kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.442292 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.450032 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.451387 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.451478 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.466941 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2tx\" (UniqueName: \"kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx\") pod \"keystone-cron-29537941-8tjbq\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.494179 4972 scope.go:117] "RemoveContainer" containerID="b1d5898eac1c4fad64bd70d62c1af61c4d98753f97bd6ca2da2729ce3c6b7021" Feb 28 11:01:00 crc kubenswrapper[4972]: I0228 11:01:00.554481 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:01 crc kubenswrapper[4972]: I0228 11:01:01.022135 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29537941-8tjbq"] Feb 28 11:01:01 crc kubenswrapper[4972]: I0228 11:01:01.787740 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29537941-8tjbq" event={"ID":"eaa3c5ff-5646-444a-b502-6052100adb2e","Type":"ContainerStarted","Data":"c4ed69374d917f2da9df6c51d1828a97cfa0c9ed91fa4556cc07531aae597c62"} Feb 28 11:01:01 crc kubenswrapper[4972]: I0228 11:01:01.807613 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29537941-8tjbq" podStartSLOduration=1.807592894 podStartE2EDuration="1.807592894s" podCreationTimestamp="2026-02-28 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 11:01:01.804747169 +0000 UTC m=+1518.716754907" watchObservedRunningTime="2026-02-28 11:01:01.807592894 +0000 UTC m=+1518.719600632" Feb 28 11:01:01 crc kubenswrapper[4972]: I0228 11:01:01.813669 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29537941-8tjbq" event={"ID":"eaa3c5ff-5646-444a-b502-6052100adb2e","Type":"ContainerStarted","Data":"9d784f1a71f298af85895062976ec1fa0b4a228713df0d5ff288ab8af3a6155e"} Feb 28 11:01:03 crc kubenswrapper[4972]: I0228 11:01:03.815400 4972 generic.go:334] "Generic (PLEG): container finished" podID="eaa3c5ff-5646-444a-b502-6052100adb2e" containerID="c4ed69374d917f2da9df6c51d1828a97cfa0c9ed91fa4556cc07531aae597c62" exitCode=0 Feb 28 11:01:03 crc kubenswrapper[4972]: I0228 11:01:03.815611 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29537941-8tjbq" event={"ID":"eaa3c5ff-5646-444a-b502-6052100adb2e","Type":"ContainerDied","Data":"c4ed69374d917f2da9df6c51d1828a97cfa0c9ed91fa4556cc07531aae597c62"} Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.146653 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.252988 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data\") pod \"eaa3c5ff-5646-444a-b502-6052100adb2e\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.253116 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys\") pod \"eaa3c5ff-5646-444a-b502-6052100adb2e\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.253247 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t2tx\" (UniqueName: \"kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx\") pod \"eaa3c5ff-5646-444a-b502-6052100adb2e\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.253326 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle\") pod \"eaa3c5ff-5646-444a-b502-6052100adb2e\" (UID: \"eaa3c5ff-5646-444a-b502-6052100adb2e\") " Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.260007 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "eaa3c5ff-5646-444a-b502-6052100adb2e" (UID: "eaa3c5ff-5646-444a-b502-6052100adb2e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.260789 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx" (OuterVolumeSpecName: "kube-api-access-9t2tx") pod "eaa3c5ff-5646-444a-b502-6052100adb2e" (UID: "eaa3c5ff-5646-444a-b502-6052100adb2e"). InnerVolumeSpecName "kube-api-access-9t2tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.294995 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaa3c5ff-5646-444a-b502-6052100adb2e" (UID: "eaa3c5ff-5646-444a-b502-6052100adb2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.331146 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data" (OuterVolumeSpecName: "config-data") pod "eaa3c5ff-5646-444a-b502-6052100adb2e" (UID: "eaa3c5ff-5646-444a-b502-6052100adb2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.355417 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t2tx\" (UniqueName: \"kubernetes.io/projected/eaa3c5ff-5646-444a-b502-6052100adb2e-kube-api-access-9t2tx\") on node \"crc\" DevicePath \"\"" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.355451 4972 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.355480 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.355493 4972 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaa3c5ff-5646-444a-b502-6052100adb2e-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.836213 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29537941-8tjbq" event={"ID":"eaa3c5ff-5646-444a-b502-6052100adb2e","Type":"ContainerDied","Data":"9d784f1a71f298af85895062976ec1fa0b4a228713df0d5ff288ab8af3a6155e"} Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.836262 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d784f1a71f298af85895062976ec1fa0b4a228713df0d5ff288ab8af3a6155e" Feb 28 11:01:05 crc kubenswrapper[4972]: I0228 11:01:05.836267 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29537941-8tjbq" Feb 28 11:01:28 crc kubenswrapper[4972]: I0228 11:01:28.890764 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:01:28 crc kubenswrapper[4972]: I0228 11:01:28.891584 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.747304 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:01:54 crc kubenswrapper[4972]: E0228 11:01:54.748399 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa3c5ff-5646-444a-b502-6052100adb2e" containerName="keystone-cron" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.748412 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa3c5ff-5646-444a-b502-6052100adb2e" containerName="keystone-cron" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.748665 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa3c5ff-5646-444a-b502-6052100adb2e" containerName="keystone-cron" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.749946 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.762888 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.852805 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.852923 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpstt\" (UniqueName: \"kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.853122 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.955591 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.955690 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpstt\" (UniqueName: \"kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.955716 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.956784 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.957026 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:54 crc kubenswrapper[4972]: I0228 11:01:54.977576 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpstt\" (UniqueName: \"kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt\") pod \"certified-operators-9zp84\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:55 crc kubenswrapper[4972]: I0228 11:01:55.084428 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:01:55 crc kubenswrapper[4972]: I0228 11:01:55.580789 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:01:56 crc kubenswrapper[4972]: I0228 11:01:56.365501 4972 generic.go:334] "Generic (PLEG): container finished" podID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerID="2edab9d2e583e6838fcb8fe20ff6caa334d93531aa90b25f9daba84fcc40be77" exitCode=0 Feb 28 11:01:56 crc kubenswrapper[4972]: I0228 11:01:56.365875 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerDied","Data":"2edab9d2e583e6838fcb8fe20ff6caa334d93531aa90b25f9daba84fcc40be77"} Feb 28 11:01:56 crc kubenswrapper[4972]: I0228 11:01:56.366116 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerStarted","Data":"3eed57667644e5fe211a0af8cf50b353968e90d42bf212412b0db2572a687af2"} Feb 28 11:01:57 crc kubenswrapper[4972]: I0228 11:01:57.376378 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerStarted","Data":"ed7ef69a7ca550148d08cac1032c582dfc051649616367e37a2fb297b89b0f37"} Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.393033 4972 generic.go:334] "Generic (PLEG): container finished" podID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerID="ed7ef69a7ca550148d08cac1032c582dfc051649616367e37a2fb297b89b0f37" exitCode=0 Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.394648 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerDied","Data":"ed7ef69a7ca550148d08cac1032c582dfc051649616367e37a2fb297b89b0f37"} Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.891581 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.892350 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.892519 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.894203 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:01:58 crc kubenswrapper[4972]: I0228 11:01:58.894345 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0" gracePeriod=600 Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.407868 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerStarted","Data":"07d757b6050d51ad44d4479488c21340e34c977b47b7888a4bb3a7395cadca6b"} Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.413104 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0" exitCode=0 Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.413143 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0"} Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.413166 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7"} Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.413188 4972 scope.go:117] "RemoveContainer" containerID="c7d19c6c589f93e6ef6e5d06a853c4145e0501875a57569d364b5b5c92c034f8" Feb 28 11:01:59 crc kubenswrapper[4972]: I0228 11:01:59.439357 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9zp84" podStartSLOduration=3.014152333 podStartE2EDuration="5.439331417s" podCreationTimestamp="2026-02-28 11:01:54 +0000 UTC" firstStartedPulling="2026-02-28 11:01:56.36855342 +0000 UTC m=+1573.280561158" lastFinishedPulling="2026-02-28 11:01:58.793732504 +0000 UTC m=+1575.705740242" observedRunningTime="2026-02-28 11:01:59.433585492 +0000 UTC m=+1576.345593230" watchObservedRunningTime="2026-02-28 11:01:59.439331417 +0000 UTC m=+1576.351339155" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.171263 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537942-sqfjd"] Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.174274 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.178019 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.178353 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.178381 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.188923 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537942-sqfjd"] Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.273851 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7rcx\" (UniqueName: \"kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx\") pod \"auto-csr-approver-29537942-sqfjd\" (UID: \"9583bd17-c0e0-405c-8ea0-f25420fae38f\") " pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.375745 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7rcx\" (UniqueName: \"kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx\") pod \"auto-csr-approver-29537942-sqfjd\" (UID: \"9583bd17-c0e0-405c-8ea0-f25420fae38f\") " pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.398452 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7rcx\" (UniqueName: \"kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx\") pod \"auto-csr-approver-29537942-sqfjd\" (UID: \"9583bd17-c0e0-405c-8ea0-f25420fae38f\") " pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.500912 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.644658 4972 scope.go:117] "RemoveContainer" containerID="235469bee88e34ed4da8757350b27b8643aeaa0bedfda01fb4d5e06d425246ee" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.679818 4972 scope.go:117] "RemoveContainer" containerID="123f1fa701904d68ab41ba1c014c5b182c037442687f66e541af8aeb25cb1959" Feb 28 11:02:00 crc kubenswrapper[4972]: I0228 11:02:00.748302 4972 scope.go:117] "RemoveContainer" containerID="bf541cc76cf02c659b3a50beeae0508bd468413e17e045507719169ba8bc0733" Feb 28 11:02:01 crc kubenswrapper[4972]: I0228 11:02:01.006180 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537942-sqfjd"] Feb 28 11:02:01 crc kubenswrapper[4972]: W0228 11:02:01.009988 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9583bd17_c0e0_405c_8ea0_f25420fae38f.slice/crio-3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077 WatchSource:0}: Error finding container 3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077: Status 404 returned error can't find the container with id 3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077 Feb 28 11:02:01 crc kubenswrapper[4972]: I0228 11:02:01.439158 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" event={"ID":"9583bd17-c0e0-405c-8ea0-f25420fae38f","Type":"ContainerStarted","Data":"3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077"} Feb 28 11:02:02 crc kubenswrapper[4972]: I0228 11:02:02.451543 4972 generic.go:334] "Generic (PLEG): container finished" podID="9583bd17-c0e0-405c-8ea0-f25420fae38f" containerID="2a6b785cf626e1e877a3e16c66489a63ab22056ba2ab3b4c39db8a6f16024f57" exitCode=0 Feb 28 11:02:02 crc kubenswrapper[4972]: I0228 11:02:02.451649 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" event={"ID":"9583bd17-c0e0-405c-8ea0-f25420fae38f","Type":"ContainerDied","Data":"2a6b785cf626e1e877a3e16c66489a63ab22056ba2ab3b4c39db8a6f16024f57"} Feb 28 11:02:03 crc kubenswrapper[4972]: I0228 11:02:03.914003 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:03 crc kubenswrapper[4972]: I0228 11:02:03.961222 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7rcx\" (UniqueName: \"kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx\") pod \"9583bd17-c0e0-405c-8ea0-f25420fae38f\" (UID: \"9583bd17-c0e0-405c-8ea0-f25420fae38f\") " Feb 28 11:02:03 crc kubenswrapper[4972]: I0228 11:02:03.975486 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx" (OuterVolumeSpecName: "kube-api-access-d7rcx") pod "9583bd17-c0e0-405c-8ea0-f25420fae38f" (UID: "9583bd17-c0e0-405c-8ea0-f25420fae38f"). InnerVolumeSpecName "kube-api-access-d7rcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:02:04 crc kubenswrapper[4972]: I0228 11:02:04.071839 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7rcx\" (UniqueName: \"kubernetes.io/projected/9583bd17-c0e0-405c-8ea0-f25420fae38f-kube-api-access-d7rcx\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:04 crc kubenswrapper[4972]: I0228 11:02:04.476472 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" event={"ID":"9583bd17-c0e0-405c-8ea0-f25420fae38f","Type":"ContainerDied","Data":"3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077"} Feb 28 11:02:04 crc kubenswrapper[4972]: I0228 11:02:04.477395 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e5b0c7fd51113094a83ab8c22d4c4b5146bb63cc2ae242cc384edd497bf6077" Feb 28 11:02:04 crc kubenswrapper[4972]: I0228 11:02:04.476529 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537942-sqfjd" Feb 28 11:02:04 crc kubenswrapper[4972]: I0228 11:02:04.997927 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537936-lwr9x"] Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.009697 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537936-lwr9x"] Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.084607 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.084695 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.161904 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.530801 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.584811 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:02:05 crc kubenswrapper[4972]: I0228 11:02:05.802433 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcceffbe-fb98-48de-a29d-51bd75e00c7a" path="/var/lib/kubelet/pods/bcceffbe-fb98-48de-a29d-51bd75e00c7a/volumes" Feb 28 11:02:07 crc kubenswrapper[4972]: I0228 11:02:07.509793 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9zp84" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="registry-server" containerID="cri-o://07d757b6050d51ad44d4479488c21340e34c977b47b7888a4bb3a7395cadca6b" gracePeriod=2 Feb 28 11:02:08 crc kubenswrapper[4972]: I0228 11:02:08.528697 4972 generic.go:334] "Generic (PLEG): container finished" podID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerID="07d757b6050d51ad44d4479488c21340e34c977b47b7888a4bb3a7395cadca6b" exitCode=0 Feb 28 11:02:08 crc kubenswrapper[4972]: I0228 11:02:08.528753 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerDied","Data":"07d757b6050d51ad44d4479488c21340e34c977b47b7888a4bb3a7395cadca6b"} Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.155934 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.191604 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content\") pod \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.191819 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities\") pod \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.191913 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpstt\" (UniqueName: \"kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt\") pod \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\" (UID: \"d4654e52-e485-41e9-91fe-6f97fb68a8ca\") " Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.192747 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities" (OuterVolumeSpecName: "utilities") pod "d4654e52-e485-41e9-91fe-6f97fb68a8ca" (UID: "d4654e52-e485-41e9-91fe-6f97fb68a8ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.198198 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt" (OuterVolumeSpecName: "kube-api-access-gpstt") pod "d4654e52-e485-41e9-91fe-6f97fb68a8ca" (UID: "d4654e52-e485-41e9-91fe-6f97fb68a8ca"). InnerVolumeSpecName "kube-api-access-gpstt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.247026 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4654e52-e485-41e9-91fe-6f97fb68a8ca" (UID: "d4654e52-e485-41e9-91fe-6f97fb68a8ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.294788 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.294830 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpstt\" (UniqueName: \"kubernetes.io/projected/d4654e52-e485-41e9-91fe-6f97fb68a8ca-kube-api-access-gpstt\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.294842 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4654e52-e485-41e9-91fe-6f97fb68a8ca-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.541489 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9zp84" event={"ID":"d4654e52-e485-41e9-91fe-6f97fb68a8ca","Type":"ContainerDied","Data":"3eed57667644e5fe211a0af8cf50b353968e90d42bf212412b0db2572a687af2"} Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.541553 4972 scope.go:117] "RemoveContainer" containerID="07d757b6050d51ad44d4479488c21340e34c977b47b7888a4bb3a7395cadca6b" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.541738 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9zp84" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.589277 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.599057 4972 scope.go:117] "RemoveContainer" containerID="ed7ef69a7ca550148d08cac1032c582dfc051649616367e37a2fb297b89b0f37" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.610118 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9zp84"] Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.633916 4972 scope.go:117] "RemoveContainer" containerID="2edab9d2e583e6838fcb8fe20ff6caa334d93531aa90b25f9daba84fcc40be77" Feb 28 11:02:09 crc kubenswrapper[4972]: I0228 11:02:09.801218 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" path="/var/lib/kubelet/pods/d4654e52-e485-41e9-91fe-6f97fb68a8ca/volumes" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.608402 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:13 crc kubenswrapper[4972]: E0228 11:02:13.610405 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="extract-utilities" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.610425 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="extract-utilities" Feb 28 11:02:13 crc kubenswrapper[4972]: E0228 11:02:13.610482 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9583bd17-c0e0-405c-8ea0-f25420fae38f" containerName="oc" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.610491 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9583bd17-c0e0-405c-8ea0-f25420fae38f" containerName="oc" Feb 28 11:02:13 crc kubenswrapper[4972]: E0228 11:02:13.610536 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="extract-content" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.610548 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="extract-content" Feb 28 11:02:13 crc kubenswrapper[4972]: E0228 11:02:13.610578 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="registry-server" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.610587 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="registry-server" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.611042 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4654e52-e485-41e9-91fe-6f97fb68a8ca" containerName="registry-server" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.611068 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9583bd17-c0e0-405c-8ea0-f25420fae38f" containerName="oc" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.614814 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.639770 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.690028 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.690437 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzj5r\" (UniqueName: \"kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.690490 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.794411 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.794593 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzj5r\" (UniqueName: \"kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.795069 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.795108 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.795069 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.818627 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzj5r\" (UniqueName: \"kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r\") pod \"community-operators-bwq7b\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:13 crc kubenswrapper[4972]: I0228 11:02:13.966852 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:14 crc kubenswrapper[4972]: I0228 11:02:14.530963 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:14 crc kubenswrapper[4972]: W0228 11:02:14.536869 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ccdfd4c_80e4_4a4f_9391_b113ce3f6574.slice/crio-6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76 WatchSource:0}: Error finding container 6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76: Status 404 returned error can't find the container with id 6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76 Feb 28 11:02:14 crc kubenswrapper[4972]: I0228 11:02:14.621732 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerStarted","Data":"6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76"} Feb 28 11:02:15 crc kubenswrapper[4972]: I0228 11:02:15.632304 4972 generic.go:334] "Generic (PLEG): container finished" podID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerID="1ee022c0cf96815f93a5a775cc7465f15b62727af031bd036979c7956af379fc" exitCode=0 Feb 28 11:02:15 crc kubenswrapper[4972]: I0228 11:02:15.632382 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerDied","Data":"1ee022c0cf96815f93a5a775cc7465f15b62727af031bd036979c7956af379fc"} Feb 28 11:02:16 crc kubenswrapper[4972]: I0228 11:02:16.740008 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerStarted","Data":"cef2dd316b1090e84773595d3417aca7dd203800e2ef4c5c89d9285b34c88807"} Feb 28 11:02:17 crc kubenswrapper[4972]: I0228 11:02:17.760712 4972 generic.go:334] "Generic (PLEG): container finished" podID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerID="cef2dd316b1090e84773595d3417aca7dd203800e2ef4c5c89d9285b34c88807" exitCode=0 Feb 28 11:02:17 crc kubenswrapper[4972]: I0228 11:02:17.760832 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerDied","Data":"cef2dd316b1090e84773595d3417aca7dd203800e2ef4c5c89d9285b34c88807"} Feb 28 11:02:18 crc kubenswrapper[4972]: I0228 11:02:18.784488 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerStarted","Data":"d60bc133fdb6b73b61a28f564e0b5a8c27daac1cad1be8152891ca20bcd5d579"} Feb 28 11:02:18 crc kubenswrapper[4972]: I0228 11:02:18.812537 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bwq7b" podStartSLOduration=3.2895104760000002 podStartE2EDuration="5.812509878s" podCreationTimestamp="2026-02-28 11:02:13 +0000 UTC" firstStartedPulling="2026-02-28 11:02:15.635038797 +0000 UTC m=+1592.547046555" lastFinishedPulling="2026-02-28 11:02:18.158038229 +0000 UTC m=+1595.070045957" observedRunningTime="2026-02-28 11:02:18.810412085 +0000 UTC m=+1595.722419823" watchObservedRunningTime="2026-02-28 11:02:18.812509878 +0000 UTC m=+1595.724517616" Feb 28 11:02:20 crc kubenswrapper[4972]: I0228 11:02:20.809556 4972 generic.go:334] "Generic (PLEG): container finished" podID="e9d8a507-2de3-4624-ad38-5d50c0c50874" containerID="342182c25f2ecb20b83bda8c3bfd6632745a41414febfb3c7f083fa12be90e68" exitCode=0 Feb 28 11:02:20 crc kubenswrapper[4972]: I0228 11:02:20.809680 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" event={"ID":"e9d8a507-2de3-4624-ad38-5d50c0c50874","Type":"ContainerDied","Data":"342182c25f2ecb20b83bda8c3bfd6632745a41414febfb3c7f083fa12be90e68"} Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.276210 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.399081 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzdbp\" (UniqueName: \"kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp\") pod \"e9d8a507-2de3-4624-ad38-5d50c0c50874\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.399277 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle\") pod \"e9d8a507-2de3-4624-ad38-5d50c0c50874\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.399375 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam\") pod \"e9d8a507-2de3-4624-ad38-5d50c0c50874\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.399446 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory\") pod \"e9d8a507-2de3-4624-ad38-5d50c0c50874\" (UID: \"e9d8a507-2de3-4624-ad38-5d50c0c50874\") " Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.407528 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp" (OuterVolumeSpecName: "kube-api-access-mzdbp") pod "e9d8a507-2de3-4624-ad38-5d50c0c50874" (UID: "e9d8a507-2de3-4624-ad38-5d50c0c50874"). InnerVolumeSpecName "kube-api-access-mzdbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.409277 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e9d8a507-2de3-4624-ad38-5d50c0c50874" (UID: "e9d8a507-2de3-4624-ad38-5d50c0c50874"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.436852 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e9d8a507-2de3-4624-ad38-5d50c0c50874" (UID: "e9d8a507-2de3-4624-ad38-5d50c0c50874"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.438103 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory" (OuterVolumeSpecName: "inventory") pod "e9d8a507-2de3-4624-ad38-5d50c0c50874" (UID: "e9d8a507-2de3-4624-ad38-5d50c0c50874"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.503561 4972 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.503605 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.503621 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d8a507-2de3-4624-ad38-5d50c0c50874-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.503632 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzdbp\" (UniqueName: \"kubernetes.io/projected/e9d8a507-2de3-4624-ad38-5d50c0c50874-kube-api-access-mzdbp\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.872050 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" event={"ID":"e9d8a507-2de3-4624-ad38-5d50c0c50874","Type":"ContainerDied","Data":"947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9"} Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.872451 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="947cb912c76a8385c2141a44c99d72a26419332453cec92d3548cfb1fea3d5b9" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.872142 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.936667 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt"] Feb 28 11:02:22 crc kubenswrapper[4972]: E0228 11:02:22.937780 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d8a507-2de3-4624-ad38-5d50c0c50874" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.937885 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d8a507-2de3-4624-ad38-5d50c0c50874" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.938367 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d8a507-2de3-4624-ad38-5d50c0c50874" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.939454 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.943281 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.943580 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.945979 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.946638 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:02:22 crc kubenswrapper[4972]: I0228 11:02:22.964779 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt"] Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.042719 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.042770 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmv2x\" (UniqueName: \"kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.042950 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.145049 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.145109 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmv2x\" (UniqueName: \"kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.145223 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.151569 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.153635 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.166708 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmv2x\" (UniqueName: \"kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.263601 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.833424 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt"] Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.907942 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" event={"ID":"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3","Type":"ContainerStarted","Data":"00074e4227ead138a99b41c4e711cd99f1671e30e450f262f0b79e4c26bd9bef"} Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.967631 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:23 crc kubenswrapper[4972]: I0228 11:02:23.967727 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:24 crc kubenswrapper[4972]: I0228 11:02:24.030751 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:24 crc kubenswrapper[4972]: I0228 11:02:24.917269 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" event={"ID":"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3","Type":"ContainerStarted","Data":"65917421612caa97c764d797b8a1ca2d002571f845cd5b9860706034af3bd07e"} Feb 28 11:02:24 crc kubenswrapper[4972]: I0228 11:02:24.945387 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" podStartSLOduration=2.200909114 podStartE2EDuration="2.945357611s" podCreationTimestamp="2026-02-28 11:02:22 +0000 UTC" firstStartedPulling="2026-02-28 11:02:23.840763291 +0000 UTC m=+1600.752771029" lastFinishedPulling="2026-02-28 11:02:24.585211788 +0000 UTC m=+1601.497219526" observedRunningTime="2026-02-28 11:02:24.940705763 +0000 UTC m=+1601.852713501" watchObservedRunningTime="2026-02-28 11:02:24.945357611 +0000 UTC m=+1601.857365349" Feb 28 11:02:24 crc kubenswrapper[4972]: I0228 11:02:24.976844 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:25 crc kubenswrapper[4972]: I0228 11:02:25.273100 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:26 crc kubenswrapper[4972]: I0228 11:02:26.945928 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bwq7b" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="registry-server" containerID="cri-o://d60bc133fdb6b73b61a28f564e0b5a8c27daac1cad1be8152891ca20bcd5d579" gracePeriod=2 Feb 28 11:02:27 crc kubenswrapper[4972]: I0228 11:02:27.958212 4972 generic.go:334] "Generic (PLEG): container finished" podID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerID="d60bc133fdb6b73b61a28f564e0b5a8c27daac1cad1be8152891ca20bcd5d579" exitCode=0 Feb 28 11:02:27 crc kubenswrapper[4972]: I0228 11:02:27.958300 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerDied","Data":"d60bc133fdb6b73b61a28f564e0b5a8c27daac1cad1be8152891ca20bcd5d579"} Feb 28 11:02:27 crc kubenswrapper[4972]: I0228 11:02:27.958673 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwq7b" event={"ID":"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574","Type":"ContainerDied","Data":"6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76"} Feb 28 11:02:27 crc kubenswrapper[4972]: I0228 11:02:27.958694 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6763f4a0a396c62fad25e236f0dbcea6940d8f5f73eff9a86891238a28423a76" Feb 28 11:02:27 crc kubenswrapper[4972]: I0228 11:02:27.996334 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.151065 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content\") pod \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.151433 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzj5r\" (UniqueName: \"kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r\") pod \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.151483 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities\") pod \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\" (UID: \"8ccdfd4c-80e4-4a4f-9391-b113ce3f6574\") " Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.153094 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities" (OuterVolumeSpecName: "utilities") pod "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" (UID: "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.160922 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r" (OuterVolumeSpecName: "kube-api-access-gzj5r") pod "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" (UID: "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574"). InnerVolumeSpecName "kube-api-access-gzj5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.205449 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" (UID: "8ccdfd4c-80e4-4a4f-9391-b113ce3f6574"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.254369 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.254415 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzj5r\" (UniqueName: \"kubernetes.io/projected/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-kube-api-access-gzj5r\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.254430 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:02:28 crc kubenswrapper[4972]: I0228 11:02:28.968597 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwq7b" Feb 28 11:02:29 crc kubenswrapper[4972]: I0228 11:02:29.016841 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:29 crc kubenswrapper[4972]: I0228 11:02:29.031350 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bwq7b"] Feb 28 11:02:29 crc kubenswrapper[4972]: I0228 11:02:29.802301 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" path="/var/lib/kubelet/pods/8ccdfd4c-80e4-4a4f-9391-b113ce3f6574/volumes" Feb 28 11:03:00 crc kubenswrapper[4972]: I0228 11:03:00.878305 4972 scope.go:117] "RemoveContainer" containerID="3d06d7a543157f8e61090743f82cc4472b8d43ab796062e2f9f7d939df6f13be" Feb 28 11:03:00 crc kubenswrapper[4972]: I0228 11:03:00.901812 4972 scope.go:117] "RemoveContainer" containerID="ea4de5fe653a69d74be9cba96ff851fee729779e9a8925a71abdc1af2532ebd2" Feb 28 11:03:00 crc kubenswrapper[4972]: I0228 11:03:00.955638 4972 scope.go:117] "RemoveContainer" containerID="567e79466eff3c51dff9964630c86e829e206b558c26a76c56c7d7d4336a075c" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.586039 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:04 crc kubenswrapper[4972]: E0228 11:03:04.587605 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="extract-utilities" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.587621 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="extract-utilities" Feb 28 11:03:04 crc kubenswrapper[4972]: E0228 11:03:04.587640 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="registry-server" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.587646 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="registry-server" Feb 28 11:03:04 crc kubenswrapper[4972]: E0228 11:03:04.587672 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="extract-content" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.587679 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="extract-content" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.587857 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ccdfd4c-80e4-4a4f-9391-b113ce3f6574" containerName="registry-server" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.594077 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.600889 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.767727 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.768250 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf7gx\" (UniqueName: \"kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.768870 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.871537 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.871645 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.871768 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf7gx\" (UniqueName: \"kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.872274 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.873013 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.897216 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf7gx\" (UniqueName: \"kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx\") pod \"redhat-marketplace-b9nrj\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:04 crc kubenswrapper[4972]: I0228 11:03:04.923045 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:05 crc kubenswrapper[4972]: I0228 11:03:05.421508 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:06 crc kubenswrapper[4972]: I0228 11:03:06.379285 4972 generic.go:334] "Generic (PLEG): container finished" podID="d0e97586-f506-4d36-bff6-67edee1414ed" containerID="b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808" exitCode=0 Feb 28 11:03:06 crc kubenswrapper[4972]: I0228 11:03:06.379799 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerDied","Data":"b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808"} Feb 28 11:03:06 crc kubenswrapper[4972]: I0228 11:03:06.379837 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerStarted","Data":"c6621da9796d4f55386f30cce3746698fc06695cacdc4e2d5c07049dd719df64"} Feb 28 11:03:07 crc kubenswrapper[4972]: I0228 11:03:07.392388 4972 generic.go:334] "Generic (PLEG): container finished" podID="d0e97586-f506-4d36-bff6-67edee1414ed" containerID="a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8" exitCode=0 Feb 28 11:03:07 crc kubenswrapper[4972]: I0228 11:03:07.392505 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerDied","Data":"a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8"} Feb 28 11:03:08 crc kubenswrapper[4972]: I0228 11:03:08.407674 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerStarted","Data":"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a"} Feb 28 11:03:08 crc kubenswrapper[4972]: I0228 11:03:08.439315 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b9nrj" podStartSLOduration=3.028553149 podStartE2EDuration="4.439289715s" podCreationTimestamp="2026-02-28 11:03:04 +0000 UTC" firstStartedPulling="2026-02-28 11:03:06.382246567 +0000 UTC m=+1643.294254305" lastFinishedPulling="2026-02-28 11:03:07.792983133 +0000 UTC m=+1644.704990871" observedRunningTime="2026-02-28 11:03:08.43200371 +0000 UTC m=+1645.344011448" watchObservedRunningTime="2026-02-28 11:03:08.439289715 +0000 UTC m=+1645.351297453" Feb 28 11:03:11 crc kubenswrapper[4972]: I0228 11:03:11.056357 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hnx6d"] Feb 28 11:03:11 crc kubenswrapper[4972]: I0228 11:03:11.070990 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hnx6d"] Feb 28 11:03:11 crc kubenswrapper[4972]: I0228 11:03:11.087258 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-cd49-account-create-update-q6sc6"] Feb 28 11:03:11 crc kubenswrapper[4972]: I0228 11:03:11.093150 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-cd49-account-create-update-q6sc6"] Feb 28 11:03:12 crc kubenswrapper[4972]: I0228 11:03:12.324381 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="600c7709-17f1-46b2-9850-673f80b46460" path="/var/lib/kubelet/pods/600c7709-17f1-46b2-9850-673f80b46460/volumes" Feb 28 11:03:12 crc kubenswrapper[4972]: I0228 11:03:12.325328 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d52cc1-c832-4ae9-929d-06db453ce86d" path="/var/lib/kubelet/pods/b8d52cc1-c832-4ae9-929d-06db453ce86d/volumes" Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.050643 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-36c3-account-create-update-hfvn9"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.063750 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-sf2wr"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.081769 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g9g7s"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.095939 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-sf2wr"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.106558 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-55bc-account-create-update-6c2kx"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.115060 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g9g7s"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.123938 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-36c3-account-create-update-hfvn9"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.134983 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-55bc-account-create-update-6c2kx"] Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.803627 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc31f7d-a528-4e8b-b7b2-015dce384926" path="/var/lib/kubelet/pods/3bc31f7d-a528-4e8b-b7b2-015dce384926/volumes" Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.804882 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467c3a6a-1854-4faa-89dd-b925b4790534" path="/var/lib/kubelet/pods/467c3a6a-1854-4faa-89dd-b925b4790534/volumes" Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.805992 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978045d2-8316-4ee1-b850-56ca45506d54" path="/var/lib/kubelet/pods/978045d2-8316-4ee1-b850-56ca45506d54/volumes" Feb 28 11:03:13 crc kubenswrapper[4972]: I0228 11:03:13.807107 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f36cf3b0-e583-4304-8e26-3142c89b8dd3" path="/var/lib/kubelet/pods/f36cf3b0-e583-4304-8e26-3142c89b8dd3/volumes" Feb 28 11:03:14 crc kubenswrapper[4972]: I0228 11:03:14.923843 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:14 crc kubenswrapper[4972]: I0228 11:03:14.924271 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:14 crc kubenswrapper[4972]: I0228 11:03:14.986832 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:15 crc kubenswrapper[4972]: I0228 11:03:15.383189 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:15 crc kubenswrapper[4972]: I0228 11:03:15.434810 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.345820 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b9nrj" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="registry-server" containerID="cri-o://7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a" gracePeriod=2 Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.871948 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.935291 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf7gx\" (UniqueName: \"kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx\") pod \"d0e97586-f506-4d36-bff6-67edee1414ed\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.935427 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content\") pod \"d0e97586-f506-4d36-bff6-67edee1414ed\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.935498 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities\") pod \"d0e97586-f506-4d36-bff6-67edee1414ed\" (UID: \"d0e97586-f506-4d36-bff6-67edee1414ed\") " Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.936417 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities" (OuterVolumeSpecName: "utilities") pod "d0e97586-f506-4d36-bff6-67edee1414ed" (UID: "d0e97586-f506-4d36-bff6-67edee1414ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.945134 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx" (OuterVolumeSpecName: "kube-api-access-hf7gx") pod "d0e97586-f506-4d36-bff6-67edee1414ed" (UID: "d0e97586-f506-4d36-bff6-67edee1414ed"). InnerVolumeSpecName "kube-api-access-hf7gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:03:17 crc kubenswrapper[4972]: I0228 11:03:17.964808 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0e97586-f506-4d36-bff6-67edee1414ed" (UID: "d0e97586-f506-4d36-bff6-67edee1414ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.038179 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf7gx\" (UniqueName: \"kubernetes.io/projected/d0e97586-f506-4d36-bff6-67edee1414ed-kube-api-access-hf7gx\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.038224 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.038235 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e97586-f506-4d36-bff6-67edee1414ed-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.367355 4972 generic.go:334] "Generic (PLEG): container finished" podID="d0e97586-f506-4d36-bff6-67edee1414ed" containerID="7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a" exitCode=0 Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.367469 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerDied","Data":"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a"} Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.367529 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9nrj" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.367554 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9nrj" event={"ID":"d0e97586-f506-4d36-bff6-67edee1414ed","Type":"ContainerDied","Data":"c6621da9796d4f55386f30cce3746698fc06695cacdc4e2d5c07049dd719df64"} Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.367583 4972 scope.go:117] "RemoveContainer" containerID="7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.413071 4972 scope.go:117] "RemoveContainer" containerID="a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.415986 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.428429 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9nrj"] Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.461745 4972 scope.go:117] "RemoveContainer" containerID="b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.493548 4972 scope.go:117] "RemoveContainer" containerID="7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a" Feb 28 11:03:18 crc kubenswrapper[4972]: E0228 11:03:18.494140 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a\": container with ID starting with 7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a not found: ID does not exist" containerID="7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.494219 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a"} err="failed to get container status \"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a\": rpc error: code = NotFound desc = could not find container \"7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a\": container with ID starting with 7d42b366792889f27747e292bdf010d2a41b817f9f866c9330c198c606ff759a not found: ID does not exist" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.494269 4972 scope.go:117] "RemoveContainer" containerID="a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8" Feb 28 11:03:18 crc kubenswrapper[4972]: E0228 11:03:18.494694 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8\": container with ID starting with a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8 not found: ID does not exist" containerID="a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.494738 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8"} err="failed to get container status \"a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8\": rpc error: code = NotFound desc = could not find container \"a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8\": container with ID starting with a97b0c95ca405a491b71a236657fc1382688555b15cc03ecebbf4811065388f8 not found: ID does not exist" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.494764 4972 scope.go:117] "RemoveContainer" containerID="b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808" Feb 28 11:03:18 crc kubenswrapper[4972]: E0228 11:03:18.495207 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808\": container with ID starting with b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808 not found: ID does not exist" containerID="b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808" Feb 28 11:03:18 crc kubenswrapper[4972]: I0228 11:03:18.495248 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808"} err="failed to get container status \"b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808\": rpc error: code = NotFound desc = could not find container \"b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808\": container with ID starting with b67160abdb3f52e9283c2fd486f9c98f265c522c0a62f8562283a57124a9b808 not found: ID does not exist" Feb 28 11:03:19 crc kubenswrapper[4972]: I0228 11:03:19.799900 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" path="/var/lib/kubelet/pods/d0e97586-f506-4d36-bff6-67edee1414ed/volumes" Feb 28 11:03:31 crc kubenswrapper[4972]: I0228 11:03:31.028341 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-5kccl"] Feb 28 11:03:31 crc kubenswrapper[4972]: I0228 11:03:31.038330 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-5kccl"] Feb 28 11:03:31 crc kubenswrapper[4972]: I0228 11:03:31.810237 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be112da-979b-4950-abb5-9d56478eefde" path="/var/lib/kubelet/pods/8be112da-979b-4950-abb5-9d56478eefde/volumes" Feb 28 11:03:41 crc kubenswrapper[4972]: I0228 11:03:41.054563 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jj7n2"] Feb 28 11:03:41 crc kubenswrapper[4972]: I0228 11:03:41.070998 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jj7n2"] Feb 28 11:03:41 crc kubenswrapper[4972]: I0228 11:03:41.826769 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c9ff65-59aa-43d9-8898-781c305625ef" path="/var/lib/kubelet/pods/15c9ff65-59aa-43d9-8898-781c305625ef/volumes" Feb 28 11:03:50 crc kubenswrapper[4972]: I0228 11:03:50.051002 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-r4c8b"] Feb 28 11:03:50 crc kubenswrapper[4972]: I0228 11:03:50.060859 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-r4c8b"] Feb 28 11:03:50 crc kubenswrapper[4972]: I0228 11:03:50.078432 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-091e-account-create-update-dhvz5"] Feb 28 11:03:50 crc kubenswrapper[4972]: I0228 11:03:50.092234 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-091e-account-create-update-dhvz5"] Feb 28 11:03:51 crc kubenswrapper[4972]: I0228 11:03:51.802261 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac57c57-313c-40e9-8118-8b2bd77be36e" path="/var/lib/kubelet/pods/0ac57c57-313c-40e9-8118-8b2bd77be36e/volumes" Feb 28 11:03:51 crc kubenswrapper[4972]: I0228 11:03:51.803057 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2491fea7-baf2-4aec-bbe6-819143abab82" path="/var/lib/kubelet/pods/2491fea7-baf2-4aec-bbe6-819143abab82/volumes" Feb 28 11:03:52 crc kubenswrapper[4972]: I0228 11:03:52.727568 4972 generic.go:334] "Generic (PLEG): container finished" podID="c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" containerID="65917421612caa97c764d797b8a1ca2d002571f845cd5b9860706034af3bd07e" exitCode=0 Feb 28 11:03:52 crc kubenswrapper[4972]: I0228 11:03:52.727692 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" event={"ID":"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3","Type":"ContainerDied","Data":"65917421612caa97c764d797b8a1ca2d002571f845cd5b9860706034af3bd07e"} Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.048532 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d4t8j"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.062243 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wbvzd"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.073747 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c659-account-create-update-xk6mx"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.081717 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-61a1-account-create-update-xzhhm"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.089077 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wbvzd"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.096950 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-61a1-account-create-update-xzhhm"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.105684 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d4t8j"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.113407 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c659-account-create-update-xk6mx"] Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.808622 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="236d1bdf-c20a-48b4-a1a2-16fd2f47910b" path="/var/lib/kubelet/pods/236d1bdf-c20a-48b4-a1a2-16fd2f47910b/volumes" Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.809827 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2760f9-f661-4ec8-8c72-f975c3f21390" path="/var/lib/kubelet/pods/2e2760f9-f661-4ec8-8c72-f975c3f21390/volumes" Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.810754 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad6d67a2-7a71-4514-9752-1b319503c1a5" path="/var/lib/kubelet/pods/ad6d67a2-7a71-4514-9752-1b319503c1a5/volumes" Feb 28 11:03:53 crc kubenswrapper[4972]: I0228 11:03:53.811631 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb0a233f-72b9-4146-b40c-a4a91620d99f" path="/var/lib/kubelet/pods/eb0a233f-72b9-4146-b40c-a4a91620d99f/volumes" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.227482 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.326537 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam\") pod \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.326655 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmv2x\" (UniqueName: \"kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x\") pod \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.326864 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory\") pod \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\" (UID: \"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3\") " Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.363635 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x" (OuterVolumeSpecName: "kube-api-access-rmv2x") pod "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" (UID: "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3"). InnerVolumeSpecName "kube-api-access-rmv2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.377872 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory" (OuterVolumeSpecName: "inventory") pod "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" (UID: "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.431279 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.431336 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmv2x\" (UniqueName: \"kubernetes.io/projected/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-kube-api-access-rmv2x\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.507778 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" (UID: "c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.535330 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.752592 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" event={"ID":"c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3","Type":"ContainerDied","Data":"00074e4227ead138a99b41c4e711cd99f1671e30e450f262f0b79e4c26bd9bef"} Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.753210 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00074e4227ead138a99b41c4e711cd99f1671e30e450f262f0b79e4c26bd9bef" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.752698 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.856873 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz"] Feb 28 11:03:54 crc kubenswrapper[4972]: E0228 11:03:54.857516 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.857548 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 28 11:03:54 crc kubenswrapper[4972]: E0228 11:03:54.857588 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="extract-content" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.857605 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="extract-content" Feb 28 11:03:54 crc kubenswrapper[4972]: E0228 11:03:54.857624 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="registry-server" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.857677 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="registry-server" Feb 28 11:03:54 crc kubenswrapper[4972]: E0228 11:03:54.857709 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="extract-utilities" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.857721 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="extract-utilities" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.858034 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.858084 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e97586-f506-4d36-bff6-67edee1414ed" containerName="registry-server" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.859174 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.863002 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.864658 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.866729 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.872501 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz"] Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.873508 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.944763 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.945150 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:54 crc kubenswrapper[4972]: I0228 11:03:54.945330 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbgll\" (UniqueName: \"kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.047545 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.047633 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.047707 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbgll\" (UniqueName: \"kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.052301 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.053128 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.067219 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbgll\" (UniqueName: \"kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.179304 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:03:55 crc kubenswrapper[4972]: I0228 11:03:55.827277 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz"] Feb 28 11:03:56 crc kubenswrapper[4972]: I0228 11:03:56.777706 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" event={"ID":"51ff636b-adde-4a49-a96d-bf1275e5b200","Type":"ContainerStarted","Data":"f917f57e2998725085212a94374b30d207ec59907601303d098408f499d263d8"} Feb 28 11:03:56 crc kubenswrapper[4972]: I0228 11:03:56.778393 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" event={"ID":"51ff636b-adde-4a49-a96d-bf1275e5b200","Type":"ContainerStarted","Data":"11ce75ec4cb485ddd5871a9393d249fb99aeb0d43433c9914d52bbcf384b25a1"} Feb 28 11:03:56 crc kubenswrapper[4972]: I0228 11:03:56.806120 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" podStartSLOduration=2.407215391 podStartE2EDuration="2.806101676s" podCreationTimestamp="2026-02-28 11:03:54 +0000 UTC" firstStartedPulling="2026-02-28 11:03:55.834357511 +0000 UTC m=+1692.746365249" lastFinishedPulling="2026-02-28 11:03:56.233243796 +0000 UTC m=+1693.145251534" observedRunningTime="2026-02-28 11:03:56.800163556 +0000 UTC m=+1693.712171294" watchObservedRunningTime="2026-02-28 11:03:56.806101676 +0000 UTC m=+1693.718109414" Feb 28 11:03:58 crc kubenswrapper[4972]: I0228 11:03:58.032546 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-w8vmt"] Feb 28 11:03:58 crc kubenswrapper[4972]: I0228 11:03:58.042072 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-w8vmt"] Feb 28 11:03:59 crc kubenswrapper[4972]: I0228 11:03:59.799807 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab7888a-c6a9-42aa-aac0-0c57d67f8613" path="/var/lib/kubelet/pods/dab7888a-c6a9-42aa-aac0-0c57d67f8613/volumes" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.143987 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537944-s7lj4"] Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.146094 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.153507 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.154169 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.154323 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.157281 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537944-s7lj4"] Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.164699 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp8ll\" (UniqueName: \"kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll\") pod \"auto-csr-approver-29537944-s7lj4\" (UID: \"e3864232-cd37-48db-a3fc-8e80a05f2312\") " pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.267346 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp8ll\" (UniqueName: \"kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll\") pod \"auto-csr-approver-29537944-s7lj4\" (UID: \"e3864232-cd37-48db-a3fc-8e80a05f2312\") " pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.304137 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp8ll\" (UniqueName: \"kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll\") pod \"auto-csr-approver-29537944-s7lj4\" (UID: \"e3864232-cd37-48db-a3fc-8e80a05f2312\") " pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:00 crc kubenswrapper[4972]: I0228 11:04:00.510226 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.001402 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537944-s7lj4"] Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.032149 4972 scope.go:117] "RemoveContainer" containerID="ff9df8f9887f1c72ffc34ff7165bbe807ceae4ff7ded130002b308c5548457f6" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.066222 4972 scope.go:117] "RemoveContainer" containerID="70f9e4a1543805bd5cca39ced121299e36f9dfd77a29a766b3b68e53ed0c9f6d" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.118126 4972 scope.go:117] "RemoveContainer" containerID="54fc94b3bd71a56f3afd5d3450540a99e13691150c7b35200c05d5246e06f92d" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.144150 4972 scope.go:117] "RemoveContainer" containerID="746c616f6ebd943a942c5c492746e775d718d8235863c502d9874665f4e81f57" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.167579 4972 scope.go:117] "RemoveContainer" containerID="ad5fac50fe00d317f996a5955886ef02ddafa83beeeb16c0e1fbcb55cf2a4cd1" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.194763 4972 scope.go:117] "RemoveContainer" containerID="5bb2e596a390c700583fdc08ea54a8defd23e9f97018d807159e54539a4a53b5" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.221115 4972 scope.go:117] "RemoveContainer" containerID="1fe7677bd53e82f2016a463899133dd11466188d9b258c5dddb71db750f6eaa9" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.249116 4972 scope.go:117] "RemoveContainer" containerID="5edf0ca16a566aece578fb843af03380f03652784ca7d0361164ca9041a26ef6" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.299393 4972 scope.go:117] "RemoveContainer" containerID="1e5e985b8a443f1d6ce01ab7e6442e29cedf58830c690417ebf19a2a45c7a744" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.382586 4972 scope.go:117] "RemoveContainer" containerID="7b4714b3e7338e8b0db3a5504f19e33365667babe0e2ddb246b544e7d3c969fb" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.450405 4972 scope.go:117] "RemoveContainer" containerID="e5e7631c598d66ce73b6cc535a6865739386b193898a6494adacbe7486fc0499" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.510656 4972 scope.go:117] "RemoveContainer" containerID="ec0cd257659fb427d608ecec5320e080a5326eca92a0d9728b42d849a27d20cf" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.562596 4972 scope.go:117] "RemoveContainer" containerID="8510522ce294b555b4e04accffd1631d6389767666e8ef127833aad14e6458e6" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.629703 4972 scope.go:117] "RemoveContainer" containerID="ecfc1dbe52e293e1adcbfe7fd0859431e4a7bde2179f20328be6a02abba47fc7" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.655393 4972 scope.go:117] "RemoveContainer" containerID="5f4d45661f3ab9791e14fd2412c98b82bf5e8a8504189159deccd97cdd72a34d" Feb 28 11:04:01 crc kubenswrapper[4972]: I0228 11:04:01.850883 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" event={"ID":"e3864232-cd37-48db-a3fc-8e80a05f2312","Type":"ContainerStarted","Data":"dc076b9d7782dff31192a29c79b1e0ed7282609c885eada1d8f5597f13ed57f6"} Feb 28 11:04:02 crc kubenswrapper[4972]: I0228 11:04:02.897978 4972 generic.go:334] "Generic (PLEG): container finished" podID="e3864232-cd37-48db-a3fc-8e80a05f2312" containerID="a196592e1e7d7382dad1f7d31cd0377ae1ddc05bed19728be064208108262785" exitCode=0 Feb 28 11:04:02 crc kubenswrapper[4972]: I0228 11:04:02.898111 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" event={"ID":"e3864232-cd37-48db-a3fc-8e80a05f2312","Type":"ContainerDied","Data":"a196592e1e7d7382dad1f7d31cd0377ae1ddc05bed19728be064208108262785"} Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.330515 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.384979 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp8ll\" (UniqueName: \"kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll\") pod \"e3864232-cd37-48db-a3fc-8e80a05f2312\" (UID: \"e3864232-cd37-48db-a3fc-8e80a05f2312\") " Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.411743 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll" (OuterVolumeSpecName: "kube-api-access-dp8ll") pod "e3864232-cd37-48db-a3fc-8e80a05f2312" (UID: "e3864232-cd37-48db-a3fc-8e80a05f2312"). InnerVolumeSpecName "kube-api-access-dp8ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.488327 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp8ll\" (UniqueName: \"kubernetes.io/projected/e3864232-cd37-48db-a3fc-8e80a05f2312-kube-api-access-dp8ll\") on node \"crc\" DevicePath \"\"" Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.926244 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" event={"ID":"e3864232-cd37-48db-a3fc-8e80a05f2312","Type":"ContainerDied","Data":"dc076b9d7782dff31192a29c79b1e0ed7282609c885eada1d8f5597f13ed57f6"} Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.926829 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc076b9d7782dff31192a29c79b1e0ed7282609c885eada1d8f5597f13ed57f6" Feb 28 11:04:04 crc kubenswrapper[4972]: I0228 11:04:04.926728 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537944-s7lj4" Feb 28 11:04:05 crc kubenswrapper[4972]: I0228 11:04:05.427597 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537938-lhgkx"] Feb 28 11:04:05 crc kubenswrapper[4972]: I0228 11:04:05.438413 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537938-lhgkx"] Feb 28 11:04:05 crc kubenswrapper[4972]: I0228 11:04:05.809699 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f217308b-0a5b-4723-bf3f-b963474cbdbb" path="/var/lib/kubelet/pods/f217308b-0a5b-4723-bf3f-b963474cbdbb/volumes" Feb 28 11:04:28 crc kubenswrapper[4972]: I0228 11:04:28.891374 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:04:28 crc kubenswrapper[4972]: I0228 11:04:28.892199 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:04:36 crc kubenswrapper[4972]: I0228 11:04:36.055796 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vppxs"] Feb 28 11:04:36 crc kubenswrapper[4972]: I0228 11:04:36.068304 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vppxs"] Feb 28 11:04:37 crc kubenswrapper[4972]: I0228 11:04:37.801366 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9e4339-b034-40f0-a1f0-04229463c846" path="/var/lib/kubelet/pods/ef9e4339-b034-40f0-a1f0-04229463c846/volumes" Feb 28 11:04:38 crc kubenswrapper[4972]: I0228 11:04:38.029690 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4kmth"] Feb 28 11:04:38 crc kubenswrapper[4972]: I0228 11:04:38.039051 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4kmth"] Feb 28 11:04:39 crc kubenswrapper[4972]: I0228 11:04:39.808733 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc233b8-4780-4e84-a0b6-dd5fe2289e16" path="/var/lib/kubelet/pods/7dc233b8-4780-4e84-a0b6-dd5fe2289e16/volumes" Feb 28 11:04:47 crc kubenswrapper[4972]: I0228 11:04:47.050237 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wwvfq"] Feb 28 11:04:47 crc kubenswrapper[4972]: I0228 11:04:47.060018 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wwvfq"] Feb 28 11:04:47 crc kubenswrapper[4972]: I0228 11:04:47.803211 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d80893f-182b-4755-afea-fd6635c700fd" path="/var/lib/kubelet/pods/4d80893f-182b-4755-afea-fd6635c700fd/volumes" Feb 28 11:04:57 crc kubenswrapper[4972]: I0228 11:04:57.037043 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qtm2c"] Feb 28 11:04:57 crc kubenswrapper[4972]: I0228 11:04:57.044423 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qtm2c"] Feb 28 11:04:57 crc kubenswrapper[4972]: I0228 11:04:57.807603 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa1673c6-2e2d-496f-b152-ca9df91df6d7" path="/var/lib/kubelet/pods/aa1673c6-2e2d-496f-b152-ca9df91df6d7/volumes" Feb 28 11:04:58 crc kubenswrapper[4972]: I0228 11:04:58.040924 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-llqn9"] Feb 28 11:04:58 crc kubenswrapper[4972]: I0228 11:04:58.056248 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-llqn9"] Feb 28 11:04:58 crc kubenswrapper[4972]: I0228 11:04:58.890876 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:04:58 crc kubenswrapper[4972]: I0228 11:04:58.891341 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:04:59 crc kubenswrapper[4972]: I0228 11:04:59.803573 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd35ce5-02ce-4909-89c5-6dfce27e50a1" path="/var/lib/kubelet/pods/edd35ce5-02ce-4909-89c5-6dfce27e50a1/volumes" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.036009 4972 scope.go:117] "RemoveContainer" containerID="db4b28d05668240068c901054c9b092d52c2356484af49604f39386d16efdecf" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.088199 4972 scope.go:117] "RemoveContainer" containerID="23096be105c71aa1516e708d9e7943de7c044738933eaf0c321ef445e9fd6661" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.162082 4972 scope.go:117] "RemoveContainer" containerID="d698e3cd64e4a29016b5708ec7fd450c9907acdd34c11aa14d442451c324ed0a" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.196532 4972 scope.go:117] "RemoveContainer" containerID="39215e36b6acf94dfdcb4712e55fdad5b1116ed4b55c3f2229e2c570a7079184" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.245372 4972 scope.go:117] "RemoveContainer" containerID="cfda70b6d18e7e219069a3a418223bc19685b3df853a742358b5ed6c6b111ea0" Feb 28 11:05:02 crc kubenswrapper[4972]: I0228 11:05:02.295809 4972 scope.go:117] "RemoveContainer" containerID="e397650d7d4ff9605e399e9aa184c180ebf09875d4634e14ffedd73ac5d41e04" Feb 28 11:05:05 crc kubenswrapper[4972]: I0228 11:05:05.562748 4972 generic.go:334] "Generic (PLEG): container finished" podID="51ff636b-adde-4a49-a96d-bf1275e5b200" containerID="f917f57e2998725085212a94374b30d207ec59907601303d098408f499d263d8" exitCode=0 Feb 28 11:05:05 crc kubenswrapper[4972]: I0228 11:05:05.562809 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" event={"ID":"51ff636b-adde-4a49-a96d-bf1275e5b200","Type":"ContainerDied","Data":"f917f57e2998725085212a94374b30d207ec59907601303d098408f499d263d8"} Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.099754 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.265670 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam\") pod \"51ff636b-adde-4a49-a96d-bf1275e5b200\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.266124 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory\") pod \"51ff636b-adde-4a49-a96d-bf1275e5b200\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.266548 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbgll\" (UniqueName: \"kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll\") pod \"51ff636b-adde-4a49-a96d-bf1275e5b200\" (UID: \"51ff636b-adde-4a49-a96d-bf1275e5b200\") " Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.285386 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll" (OuterVolumeSpecName: "kube-api-access-jbgll") pod "51ff636b-adde-4a49-a96d-bf1275e5b200" (UID: "51ff636b-adde-4a49-a96d-bf1275e5b200"). InnerVolumeSpecName "kube-api-access-jbgll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.297388 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "51ff636b-adde-4a49-a96d-bf1275e5b200" (UID: "51ff636b-adde-4a49-a96d-bf1275e5b200"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.330239 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory" (OuterVolumeSpecName: "inventory") pod "51ff636b-adde-4a49-a96d-bf1275e5b200" (UID: "51ff636b-adde-4a49-a96d-bf1275e5b200"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.369947 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.370019 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51ff636b-adde-4a49-a96d-bf1275e5b200-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.370034 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbgll\" (UniqueName: \"kubernetes.io/projected/51ff636b-adde-4a49-a96d-bf1275e5b200-kube-api-access-jbgll\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.583643 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" event={"ID":"51ff636b-adde-4a49-a96d-bf1275e5b200","Type":"ContainerDied","Data":"11ce75ec4cb485ddd5871a9393d249fb99aeb0d43433c9914d52bbcf384b25a1"} Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.584065 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11ce75ec4cb485ddd5871a9393d249fb99aeb0d43433c9914d52bbcf384b25a1" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.583769 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.707243 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz"] Feb 28 11:05:07 crc kubenswrapper[4972]: E0228 11:05:07.707703 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ff636b-adde-4a49-a96d-bf1275e5b200" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.707726 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ff636b-adde-4a49-a96d-bf1275e5b200" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:07 crc kubenswrapper[4972]: E0228 11:05:07.707743 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3864232-cd37-48db-a3fc-8e80a05f2312" containerName="oc" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.707752 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3864232-cd37-48db-a3fc-8e80a05f2312" containerName="oc" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.707920 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ff636b-adde-4a49-a96d-bf1275e5b200" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.707937 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3864232-cd37-48db-a3fc-8e80a05f2312" containerName="oc" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.708641 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.712830 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.712993 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.713209 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.713379 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.730765 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz"] Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.881139 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.881525 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p49kx\" (UniqueName: \"kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.881891 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.984497 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p49kx\" (UniqueName: \"kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.984629 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.984734 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.992382 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:07 crc kubenswrapper[4972]: I0228 11:05:07.992578 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:08 crc kubenswrapper[4972]: I0228 11:05:08.007807 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p49kx\" (UniqueName: \"kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:08 crc kubenswrapper[4972]: I0228 11:05:08.039533 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:08 crc kubenswrapper[4972]: I0228 11:05:08.610314 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:05:08 crc kubenswrapper[4972]: I0228 11:05:08.611701 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz"] Feb 28 11:05:09 crc kubenswrapper[4972]: I0228 11:05:09.603724 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" event={"ID":"6436504e-d8a5-4dc9-b500-810a55ef3bbb","Type":"ContainerStarted","Data":"4a9e0fd9240dc690bfdb6499f703b003a02371a43c8efa58a9a57bc00b821b27"} Feb 28 11:05:09 crc kubenswrapper[4972]: I0228 11:05:09.604184 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" event={"ID":"6436504e-d8a5-4dc9-b500-810a55ef3bbb","Type":"ContainerStarted","Data":"cc3a0b7671d8d960444e16e3f59d95a474d72ab8e28682781257d701dc35c321"} Feb 28 11:05:09 crc kubenswrapper[4972]: I0228 11:05:09.634885 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" podStartSLOduration=2.174112517 podStartE2EDuration="2.634797467s" podCreationTimestamp="2026-02-28 11:05:07 +0000 UTC" firstStartedPulling="2026-02-28 11:05:08.610066158 +0000 UTC m=+1765.522073896" lastFinishedPulling="2026-02-28 11:05:09.070751108 +0000 UTC m=+1765.982758846" observedRunningTime="2026-02-28 11:05:09.619787224 +0000 UTC m=+1766.531795012" watchObservedRunningTime="2026-02-28 11:05:09.634797467 +0000 UTC m=+1766.546805215" Feb 28 11:05:14 crc kubenswrapper[4972]: I0228 11:05:14.661808 4972 generic.go:334] "Generic (PLEG): container finished" podID="6436504e-d8a5-4dc9-b500-810a55ef3bbb" containerID="4a9e0fd9240dc690bfdb6499f703b003a02371a43c8efa58a9a57bc00b821b27" exitCode=0 Feb 28 11:05:14 crc kubenswrapper[4972]: I0228 11:05:14.661905 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" event={"ID":"6436504e-d8a5-4dc9-b500-810a55ef3bbb","Type":"ContainerDied","Data":"4a9e0fd9240dc690bfdb6499f703b003a02371a43c8efa58a9a57bc00b821b27"} Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.078977 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.094501 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p49kx\" (UniqueName: \"kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx\") pod \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.094786 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory\") pod \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.095102 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam\") pod \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\" (UID: \"6436504e-d8a5-4dc9-b500-810a55ef3bbb\") " Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.107967 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx" (OuterVolumeSpecName: "kube-api-access-p49kx") pod "6436504e-d8a5-4dc9-b500-810a55ef3bbb" (UID: "6436504e-d8a5-4dc9-b500-810a55ef3bbb"). InnerVolumeSpecName "kube-api-access-p49kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.132506 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory" (OuterVolumeSpecName: "inventory") pod "6436504e-d8a5-4dc9-b500-810a55ef3bbb" (UID: "6436504e-d8a5-4dc9-b500-810a55ef3bbb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.145340 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6436504e-d8a5-4dc9-b500-810a55ef3bbb" (UID: "6436504e-d8a5-4dc9-b500-810a55ef3bbb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.198612 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.198674 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6436504e-d8a5-4dc9-b500-810a55ef3bbb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.198688 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p49kx\" (UniqueName: \"kubernetes.io/projected/6436504e-d8a5-4dc9-b500-810a55ef3bbb-kube-api-access-p49kx\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.695379 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" event={"ID":"6436504e-d8a5-4dc9-b500-810a55ef3bbb","Type":"ContainerDied","Data":"cc3a0b7671d8d960444e16e3f59d95a474d72ab8e28682781257d701dc35c321"} Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.695856 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc3a0b7671d8d960444e16e3f59d95a474d72ab8e28682781257d701dc35c321" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.695575 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.773276 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp"] Feb 28 11:05:16 crc kubenswrapper[4972]: E0228 11:05:16.774027 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6436504e-d8a5-4dc9-b500-810a55ef3bbb" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.774063 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="6436504e-d8a5-4dc9-b500-810a55ef3bbb" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.774288 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="6436504e-d8a5-4dc9-b500-810a55ef3bbb" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.775080 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.779848 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.779980 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.780017 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.780217 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.818811 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp"] Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.823156 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.823270 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.823299 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbstv\" (UniqueName: \"kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.926352 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.926504 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.926530 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbstv\" (UniqueName: \"kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.935400 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.936261 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:16 crc kubenswrapper[4972]: I0228 11:05:16.946451 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbstv\" (UniqueName: \"kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9kfmp\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:17 crc kubenswrapper[4972]: I0228 11:05:17.113979 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:17 crc kubenswrapper[4972]: I0228 11:05:17.515872 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp"] Feb 28 11:05:17 crc kubenswrapper[4972]: I0228 11:05:17.706506 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" event={"ID":"8a963845-f99d-488b-81c8-b13e44cc5ff2","Type":"ContainerStarted","Data":"f3ef90ac491d6fa241707aec2e403b0e160c6c52ccf35442652d3d351d387a04"} Feb 28 11:05:18 crc kubenswrapper[4972]: I0228 11:05:18.723204 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" event={"ID":"8a963845-f99d-488b-81c8-b13e44cc5ff2","Type":"ContainerStarted","Data":"19ff513a5bead8d6db935acc1f7a36d75badf7c465e8860728dd5ee95f28e33d"} Feb 28 11:05:18 crc kubenswrapper[4972]: I0228 11:05:18.752524 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" podStartSLOduration=2.316831202 podStartE2EDuration="2.752477134s" podCreationTimestamp="2026-02-28 11:05:16 +0000 UTC" firstStartedPulling="2026-02-28 11:05:17.522708806 +0000 UTC m=+1774.434716564" lastFinishedPulling="2026-02-28 11:05:17.958354718 +0000 UTC m=+1774.870362496" observedRunningTime="2026-02-28 11:05:18.74470825 +0000 UTC m=+1775.656715988" watchObservedRunningTime="2026-02-28 11:05:18.752477134 +0000 UTC m=+1775.664484882" Feb 28 11:05:28 crc kubenswrapper[4972]: I0228 11:05:28.891794 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:05:28 crc kubenswrapper[4972]: I0228 11:05:28.892784 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:05:28 crc kubenswrapper[4972]: I0228 11:05:28.892923 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:05:28 crc kubenswrapper[4972]: I0228 11:05:28.894599 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:05:28 crc kubenswrapper[4972]: I0228 11:05:28.894723 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" gracePeriod=600 Feb 28 11:05:29 crc kubenswrapper[4972]: E0228 11:05:29.045413 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:05:29 crc kubenswrapper[4972]: I0228 11:05:29.883679 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" exitCode=0 Feb 28 11:05:29 crc kubenswrapper[4972]: I0228 11:05:29.883771 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7"} Feb 28 11:05:29 crc kubenswrapper[4972]: I0228 11:05:29.884149 4972 scope.go:117] "RemoveContainer" containerID="f31f1079c99acf17a4af5348d07b4066ed77718bdb9114498eb8336ee7f1c9a0" Feb 28 11:05:29 crc kubenswrapper[4972]: I0228 11:05:29.886539 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:05:29 crc kubenswrapper[4972]: E0228 11:05:29.887384 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:05:39 crc kubenswrapper[4972]: I0228 11:05:39.064839 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-dmxzn"] Feb 28 11:05:39 crc kubenswrapper[4972]: I0228 11:05:39.081638 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-dmxzn"] Feb 28 11:05:39 crc kubenswrapper[4972]: I0228 11:05:39.802232 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9dd22f-f39a-48c5-9c92-3b77871d93aa" path="/var/lib/kubelet/pods/8c9dd22f-f39a-48c5-9c92-3b77871d93aa/volumes" Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.042961 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-qspjr"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.055759 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-53fe-account-create-update-b6lq2"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.067538 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-gt8dq"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.078753 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-53fe-account-create-update-b6lq2"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.090374 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-qspjr"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.101919 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-93d4-account-create-update-w8pl2"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.112210 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-gt8dq"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.123479 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7008-account-create-update-xtgq2"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.134444 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-93d4-account-create-update-w8pl2"] Feb 28 11:05:40 crc kubenswrapper[4972]: I0228 11:05:40.147826 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7008-account-create-update-xtgq2"] Feb 28 11:05:41 crc kubenswrapper[4972]: I0228 11:05:41.806423 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ae66a7-72f6-445b-aa8a-f6e820d86743" path="/var/lib/kubelet/pods/20ae66a7-72f6-445b-aa8a-f6e820d86743/volumes" Feb 28 11:05:41 crc kubenswrapper[4972]: I0228 11:05:41.807707 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d9ab92-ebdd-48e1-bcff-c7baf040dee3" path="/var/lib/kubelet/pods/45d9ab92-ebdd-48e1-bcff-c7baf040dee3/volumes" Feb 28 11:05:41 crc kubenswrapper[4972]: I0228 11:05:41.808439 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a679e6b9-6ad9-450f-8b9e-2d37588bfea8" path="/var/lib/kubelet/pods/a679e6b9-6ad9-450f-8b9e-2d37588bfea8/volumes" Feb 28 11:05:41 crc kubenswrapper[4972]: I0228 11:05:41.809270 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e03b775f-1119-4c4b-8661-74213244355e" path="/var/lib/kubelet/pods/e03b775f-1119-4c4b-8661-74213244355e/volumes" Feb 28 11:05:41 crc kubenswrapper[4972]: I0228 11:05:41.810746 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5fd2836-500a-432f-b537-db8ddd108097" path="/var/lib/kubelet/pods/f5fd2836-500a-432f-b537-db8ddd108097/volumes" Feb 28 11:05:43 crc kubenswrapper[4972]: I0228 11:05:43.798090 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:05:43 crc kubenswrapper[4972]: E0228 11:05:43.799572 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:05:56 crc kubenswrapper[4972]: I0228 11:05:56.796689 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:05:56 crc kubenswrapper[4972]: E0228 11:05:56.814904 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:05:58 crc kubenswrapper[4972]: I0228 11:05:58.253304 4972 generic.go:334] "Generic (PLEG): container finished" podID="8a963845-f99d-488b-81c8-b13e44cc5ff2" containerID="19ff513a5bead8d6db935acc1f7a36d75badf7c465e8860728dd5ee95f28e33d" exitCode=0 Feb 28 11:05:58 crc kubenswrapper[4972]: I0228 11:05:58.253357 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" event={"ID":"8a963845-f99d-488b-81c8-b13e44cc5ff2","Type":"ContainerDied","Data":"19ff513a5bead8d6db935acc1f7a36d75badf7c465e8860728dd5ee95f28e33d"} Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.679179 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.807560 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory\") pod \"8a963845-f99d-488b-81c8-b13e44cc5ff2\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.807896 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam\") pod \"8a963845-f99d-488b-81c8-b13e44cc5ff2\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.807935 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbstv\" (UniqueName: \"kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv\") pod \"8a963845-f99d-488b-81c8-b13e44cc5ff2\" (UID: \"8a963845-f99d-488b-81c8-b13e44cc5ff2\") " Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.812767 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv" (OuterVolumeSpecName: "kube-api-access-sbstv") pod "8a963845-f99d-488b-81c8-b13e44cc5ff2" (UID: "8a963845-f99d-488b-81c8-b13e44cc5ff2"). InnerVolumeSpecName "kube-api-access-sbstv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.839017 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8a963845-f99d-488b-81c8-b13e44cc5ff2" (UID: "8a963845-f99d-488b-81c8-b13e44cc5ff2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.839100 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory" (OuterVolumeSpecName: "inventory") pod "8a963845-f99d-488b-81c8-b13e44cc5ff2" (UID: "8a963845-f99d-488b-81c8-b13e44cc5ff2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.911961 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.912004 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbstv\" (UniqueName: \"kubernetes.io/projected/8a963845-f99d-488b-81c8-b13e44cc5ff2-kube-api-access-sbstv\") on node \"crc\" DevicePath \"\"" Feb 28 11:05:59 crc kubenswrapper[4972]: I0228 11:05:59.912020 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a963845-f99d-488b-81c8-b13e44cc5ff2-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.158776 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537946-nzd87"] Feb 28 11:06:00 crc kubenswrapper[4972]: E0228 11:06:00.159832 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a963845-f99d-488b-81c8-b13e44cc5ff2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.159860 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a963845-f99d-488b-81c8-b13e44cc5ff2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.160109 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a963845-f99d-488b-81c8-b13e44cc5ff2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.160957 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.163425 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.163654 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.163839 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.174131 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537946-nzd87"] Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.218901 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4psxp\" (UniqueName: \"kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp\") pod \"auto-csr-approver-29537946-nzd87\" (UID: \"6510a03f-c392-4cde-b9be-a0a6405cf2c3\") " pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.275076 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.275105 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9kfmp" event={"ID":"8a963845-f99d-488b-81c8-b13e44cc5ff2","Type":"ContainerDied","Data":"f3ef90ac491d6fa241707aec2e403b0e160c6c52ccf35442652d3d351d387a04"} Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.275158 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3ef90ac491d6fa241707aec2e403b0e160c6c52ccf35442652d3d351d387a04" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.335133 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4psxp\" (UniqueName: \"kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp\") pod \"auto-csr-approver-29537946-nzd87\" (UID: \"6510a03f-c392-4cde-b9be-a0a6405cf2c3\") " pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.389345 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4psxp\" (UniqueName: \"kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp\") pod \"auto-csr-approver-29537946-nzd87\" (UID: \"6510a03f-c392-4cde-b9be-a0a6405cf2c3\") " pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.396770 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh"] Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.398817 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.403803 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.403989 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.404046 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.404248 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.422279 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh"] Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.497891 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.539962 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.540148 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhkgr\" (UniqueName: \"kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.540249 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.641833 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.642426 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhkgr\" (UniqueName: \"kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.643129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.655365 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.655696 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.660819 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhkgr\" (UniqueName: \"kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q2clh\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:00 crc kubenswrapper[4972]: I0228 11:06:00.753650 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:01 crc kubenswrapper[4972]: I0228 11:06:01.017241 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537946-nzd87"] Feb 28 11:06:01 crc kubenswrapper[4972]: I0228 11:06:01.290367 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh"] Feb 28 11:06:01 crc kubenswrapper[4972]: I0228 11:06:01.293628 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537946-nzd87" event={"ID":"6510a03f-c392-4cde-b9be-a0a6405cf2c3","Type":"ContainerStarted","Data":"2b103d98fb236ae2596610de3a5bf5d46b4956c3909e0d5103f4b122e7734aad"} Feb 28 11:06:01 crc kubenswrapper[4972]: W0228 11:06:01.297581 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc25a96_f5f2_491d_b342_41bb674572ba.slice/crio-d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac WatchSource:0}: Error finding container d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac: Status 404 returned error can't find the container with id d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.305404 4972 generic.go:334] "Generic (PLEG): container finished" podID="6510a03f-c392-4cde-b9be-a0a6405cf2c3" containerID="e756f91a1e08e3413bb3da9f0e0cffad199c75d096c6a43e9bd66c912ea894e0" exitCode=0 Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.305526 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537946-nzd87" event={"ID":"6510a03f-c392-4cde-b9be-a0a6405cf2c3","Type":"ContainerDied","Data":"e756f91a1e08e3413bb3da9f0e0cffad199c75d096c6a43e9bd66c912ea894e0"} Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.309359 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" event={"ID":"5dc25a96-f5f2-491d-b342-41bb674572ba","Type":"ContainerStarted","Data":"9fe356ef68b86ba2284e8de98cbabf903adc221c79018b1ab956ff87a15ed341"} Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.309528 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" event={"ID":"5dc25a96-f5f2-491d-b342-41bb674572ba","Type":"ContainerStarted","Data":"d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac"} Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.358327 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" podStartSLOduration=1.97375627 podStartE2EDuration="2.358299636s" podCreationTimestamp="2026-02-28 11:06:00 +0000 UTC" firstStartedPulling="2026-02-28 11:06:01.299892592 +0000 UTC m=+1818.211900330" lastFinishedPulling="2026-02-28 11:06:01.684435958 +0000 UTC m=+1818.596443696" observedRunningTime="2026-02-28 11:06:02.347268762 +0000 UTC m=+1819.259276500" watchObservedRunningTime="2026-02-28 11:06:02.358299636 +0000 UTC m=+1819.270307374" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.449617 4972 scope.go:117] "RemoveContainer" containerID="dfd19676c642577cd47a922d6d5a26f16344d989e82ae13af7b8f10983f41e33" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.489542 4972 scope.go:117] "RemoveContainer" containerID="71c97183b340cff7d37bf1aab9dc2715563329ef2ba4e0e6a64ba5f365d22255" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.538603 4972 scope.go:117] "RemoveContainer" containerID="b080eaf67a6d54e7254947e162be62df77ad2b581da656755e6333f33e6b5e80" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.567228 4972 scope.go:117] "RemoveContainer" containerID="213c44d71e4f4791450c4c082e111096fa2e1ee5168aaffd8fa0d5410bb9118c" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.601102 4972 scope.go:117] "RemoveContainer" containerID="f1b696938685daeb81be165508e5ff9e604e1055bb09e9c2facc277b311948f9" Feb 28 11:06:02 crc kubenswrapper[4972]: I0228 11:06:02.644395 4972 scope.go:117] "RemoveContainer" containerID="01aaa1f5ceec02d6680959beec691308ab6f06f8d3eb32958dc21a8dfbd75f75" Feb 28 11:06:03 crc kubenswrapper[4972]: I0228 11:06:03.719500 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:03 crc kubenswrapper[4972]: I0228 11:06:03.831026 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4psxp\" (UniqueName: \"kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp\") pod \"6510a03f-c392-4cde-b9be-a0a6405cf2c3\" (UID: \"6510a03f-c392-4cde-b9be-a0a6405cf2c3\") " Feb 28 11:06:03 crc kubenswrapper[4972]: I0228 11:06:03.840711 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp" (OuterVolumeSpecName: "kube-api-access-4psxp") pod "6510a03f-c392-4cde-b9be-a0a6405cf2c3" (UID: "6510a03f-c392-4cde-b9be-a0a6405cf2c3"). InnerVolumeSpecName "kube-api-access-4psxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:06:03 crc kubenswrapper[4972]: I0228 11:06:03.933927 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4psxp\" (UniqueName: \"kubernetes.io/projected/6510a03f-c392-4cde-b9be-a0a6405cf2c3-kube-api-access-4psxp\") on node \"crc\" DevicePath \"\"" Feb 28 11:06:04 crc kubenswrapper[4972]: I0228 11:06:04.336964 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537946-nzd87" event={"ID":"6510a03f-c392-4cde-b9be-a0a6405cf2c3","Type":"ContainerDied","Data":"2b103d98fb236ae2596610de3a5bf5d46b4956c3909e0d5103f4b122e7734aad"} Feb 28 11:06:04 crc kubenswrapper[4972]: I0228 11:06:04.337038 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b103d98fb236ae2596610de3a5bf5d46b4956c3909e0d5103f4b122e7734aad" Feb 28 11:06:04 crc kubenswrapper[4972]: I0228 11:06:04.337437 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537946-nzd87" Feb 28 11:06:04 crc kubenswrapper[4972]: I0228 11:06:04.806243 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537940-mgnwl"] Feb 28 11:06:04 crc kubenswrapper[4972]: I0228 11:06:04.816726 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537940-mgnwl"] Feb 28 11:06:05 crc kubenswrapper[4972]: I0228 11:06:05.808979 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7766b68b-d801-4a78-b664-9ef12c2d30c0" path="/var/lib/kubelet/pods/7766b68b-d801-4a78-b664-9ef12c2d30c0/volumes" Feb 28 11:06:10 crc kubenswrapper[4972]: I0228 11:06:10.789787 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:06:10 crc kubenswrapper[4972]: E0228 11:06:10.790682 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:06:22 crc kubenswrapper[4972]: I0228 11:06:22.044913 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-56fjs"] Feb 28 11:06:22 crc kubenswrapper[4972]: I0228 11:06:22.054935 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-56fjs"] Feb 28 11:06:23 crc kubenswrapper[4972]: I0228 11:06:23.801508 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7caa58-92af-435f-9d46-6f63ed19a6e6" path="/var/lib/kubelet/pods/6e7caa58-92af-435f-9d46-6f63ed19a6e6/volumes" Feb 28 11:06:24 crc kubenswrapper[4972]: I0228 11:06:24.789389 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:06:24 crc kubenswrapper[4972]: E0228 11:06:24.790160 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:06:36 crc kubenswrapper[4972]: I0228 11:06:36.789794 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:06:36 crc kubenswrapper[4972]: E0228 11:06:36.807767 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:06:42 crc kubenswrapper[4972]: I0228 11:06:42.065782 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrmpd"] Feb 28 11:06:42 crc kubenswrapper[4972]: I0228 11:06:42.078086 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lrmpd"] Feb 28 11:06:43 crc kubenswrapper[4972]: I0228 11:06:43.807692 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67229e04-2f99-4419-b160-a8871c61c886" path="/var/lib/kubelet/pods/67229e04-2f99-4419-b160-a8871c61c886/volumes" Feb 28 11:06:48 crc kubenswrapper[4972]: I0228 11:06:48.039346 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-522pb"] Feb 28 11:06:48 crc kubenswrapper[4972]: I0228 11:06:48.047579 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-522pb"] Feb 28 11:06:49 crc kubenswrapper[4972]: I0228 11:06:49.789605 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:06:49 crc kubenswrapper[4972]: E0228 11:06:49.790137 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:06:49 crc kubenswrapper[4972]: I0228 11:06:49.800326 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39eca2d3-a87e-47f8-9d3b-ee07a5969701" path="/var/lib/kubelet/pods/39eca2d3-a87e-47f8-9d3b-ee07a5969701/volumes" Feb 28 11:06:53 crc kubenswrapper[4972]: I0228 11:06:53.924834 4972 generic.go:334] "Generic (PLEG): container finished" podID="5dc25a96-f5f2-491d-b342-41bb674572ba" containerID="9fe356ef68b86ba2284e8de98cbabf903adc221c79018b1ab956ff87a15ed341" exitCode=0 Feb 28 11:06:53 crc kubenswrapper[4972]: I0228 11:06:53.925056 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" event={"ID":"5dc25a96-f5f2-491d-b342-41bb674572ba","Type":"ContainerDied","Data":"9fe356ef68b86ba2284e8de98cbabf903adc221c79018b1ab956ff87a15ed341"} Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.480792 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.545568 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam\") pod \"5dc25a96-f5f2-491d-b342-41bb674572ba\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.545981 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhkgr\" (UniqueName: \"kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr\") pod \"5dc25a96-f5f2-491d-b342-41bb674572ba\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.546014 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory\") pod \"5dc25a96-f5f2-491d-b342-41bb674572ba\" (UID: \"5dc25a96-f5f2-491d-b342-41bb674572ba\") " Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.553945 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr" (OuterVolumeSpecName: "kube-api-access-jhkgr") pod "5dc25a96-f5f2-491d-b342-41bb674572ba" (UID: "5dc25a96-f5f2-491d-b342-41bb674572ba"). InnerVolumeSpecName "kube-api-access-jhkgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.619825 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory" (OuterVolumeSpecName: "inventory") pod "5dc25a96-f5f2-491d-b342-41bb674572ba" (UID: "5dc25a96-f5f2-491d-b342-41bb674572ba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.620171 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5dc25a96-f5f2-491d-b342-41bb674572ba" (UID: "5dc25a96-f5f2-491d-b342-41bb674572ba"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.648711 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.648746 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhkgr\" (UniqueName: \"kubernetes.io/projected/5dc25a96-f5f2-491d-b342-41bb674572ba-kube-api-access-jhkgr\") on node \"crc\" DevicePath \"\"" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.648759 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc25a96-f5f2-491d-b342-41bb674572ba-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.951944 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" event={"ID":"5dc25a96-f5f2-491d-b342-41bb674572ba","Type":"ContainerDied","Data":"d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac"} Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.951999 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7aa1b9875c5de3a5dc889a9618effc5fa040febf74fec762f9b4880f77e97ac" Feb 28 11:06:55 crc kubenswrapper[4972]: I0228 11:06:55.952083 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q2clh" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.075231 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nqj8m"] Feb 28 11:06:56 crc kubenswrapper[4972]: E0228 11:06:56.075671 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6510a03f-c392-4cde-b9be-a0a6405cf2c3" containerName="oc" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.075693 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="6510a03f-c392-4cde-b9be-a0a6405cf2c3" containerName="oc" Feb 28 11:06:56 crc kubenswrapper[4972]: E0228 11:06:56.075731 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc25a96-f5f2-491d-b342-41bb674572ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.075739 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc25a96-f5f2-491d-b342-41bb674572ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.075897 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc25a96-f5f2-491d-b342-41bb674572ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.075922 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="6510a03f-c392-4cde-b9be-a0a6405cf2c3" containerName="oc" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.076675 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.079568 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.080253 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.080487 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.082265 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.087399 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nqj8m"] Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.158517 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.158595 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szljj\" (UniqueName: \"kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.158746 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.261154 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.261225 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szljj\" (UniqueName: \"kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.261302 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.265408 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.271071 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.282170 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szljj\" (UniqueName: \"kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj\") pod \"ssh-known-hosts-edpm-deployment-nqj8m\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:56 crc kubenswrapper[4972]: I0228 11:06:56.420326 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:06:57 crc kubenswrapper[4972]: I0228 11:06:57.123576 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nqj8m"] Feb 28 11:06:57 crc kubenswrapper[4972]: W0228 11:06:57.130671 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41b5f77f_8955_4d82_8d40_08c398090bc7.slice/crio-d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1 WatchSource:0}: Error finding container d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1: Status 404 returned error can't find the container with id d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1 Feb 28 11:06:57 crc kubenswrapper[4972]: I0228 11:06:57.973814 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" event={"ID":"41b5f77f-8955-4d82-8d40-08c398090bc7","Type":"ContainerStarted","Data":"d9af6ed47bf04242e0582ac4e4a2e6d58570811c306fd4254668176ddefa8a41"} Feb 28 11:06:57 crc kubenswrapper[4972]: I0228 11:06:57.974566 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" event={"ID":"41b5f77f-8955-4d82-8d40-08c398090bc7","Type":"ContainerStarted","Data":"d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1"} Feb 28 11:07:01 crc kubenswrapper[4972]: I0228 11:07:01.791793 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:07:01 crc kubenswrapper[4972]: E0228 11:07:01.793551 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:07:02 crc kubenswrapper[4972]: I0228 11:07:02.845520 4972 scope.go:117] "RemoveContainer" containerID="a6cdca71b0d7d252124e9d485c53110feb4235f37dbb35cf332a1d2a62d6750b" Feb 28 11:07:02 crc kubenswrapper[4972]: I0228 11:07:02.898826 4972 scope.go:117] "RemoveContainer" containerID="c54d631349c50bf9a494cdd8ad00bcc19510126580db34ee1f024e06388ed8fc" Feb 28 11:07:02 crc kubenswrapper[4972]: I0228 11:07:02.977501 4972 scope.go:117] "RemoveContainer" containerID="7a0084846917d0aaf5389f62c60cd27ee77ce90eb5d14d810ba3f5add64801b7" Feb 28 11:07:03 crc kubenswrapper[4972]: I0228 11:07:03.030677 4972 scope.go:117] "RemoveContainer" containerID="af27305d025d2829b00a450c674e3957338dcee34e93cfec610ec7229627c73c" Feb 28 11:07:05 crc kubenswrapper[4972]: I0228 11:07:05.057777 4972 generic.go:334] "Generic (PLEG): container finished" podID="41b5f77f-8955-4d82-8d40-08c398090bc7" containerID="d9af6ed47bf04242e0582ac4e4a2e6d58570811c306fd4254668176ddefa8a41" exitCode=0 Feb 28 11:07:05 crc kubenswrapper[4972]: I0228 11:07:05.058251 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" event={"ID":"41b5f77f-8955-4d82-8d40-08c398090bc7","Type":"ContainerDied","Data":"d9af6ed47bf04242e0582ac4e4a2e6d58570811c306fd4254668176ddefa8a41"} Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.628436 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.781974 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam\") pod \"41b5f77f-8955-4d82-8d40-08c398090bc7\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.782066 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0\") pod \"41b5f77f-8955-4d82-8d40-08c398090bc7\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.782370 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szljj\" (UniqueName: \"kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj\") pod \"41b5f77f-8955-4d82-8d40-08c398090bc7\" (UID: \"41b5f77f-8955-4d82-8d40-08c398090bc7\") " Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.788688 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj" (OuterVolumeSpecName: "kube-api-access-szljj") pod "41b5f77f-8955-4d82-8d40-08c398090bc7" (UID: "41b5f77f-8955-4d82-8d40-08c398090bc7"). InnerVolumeSpecName "kube-api-access-szljj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.816926 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "41b5f77f-8955-4d82-8d40-08c398090bc7" (UID: "41b5f77f-8955-4d82-8d40-08c398090bc7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.824955 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41b5f77f-8955-4d82-8d40-08c398090bc7" (UID: "41b5f77f-8955-4d82-8d40-08c398090bc7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.885164 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szljj\" (UniqueName: \"kubernetes.io/projected/41b5f77f-8955-4d82-8d40-08c398090bc7-kube-api-access-szljj\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.885208 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:06 crc kubenswrapper[4972]: I0228 11:07:06.885223 4972 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/41b5f77f-8955-4d82-8d40-08c398090bc7-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.079210 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" event={"ID":"41b5f77f-8955-4d82-8d40-08c398090bc7","Type":"ContainerDied","Data":"d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1"} Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.079258 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d290d59a2c7f1dd8ea0faa8102d8d4b17694d7b7cb20a9aff353f235ec7044b1" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.079312 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nqj8m" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.172664 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4"] Feb 28 11:07:07 crc kubenswrapper[4972]: E0228 11:07:07.173275 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b5f77f-8955-4d82-8d40-08c398090bc7" containerName="ssh-known-hosts-edpm-deployment" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.173308 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b5f77f-8955-4d82-8d40-08c398090bc7" containerName="ssh-known-hosts-edpm-deployment" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.173665 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b5f77f-8955-4d82-8d40-08c398090bc7" containerName="ssh-known-hosts-edpm-deployment" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.174743 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.177113 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.178353 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.181392 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.181608 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.182765 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4"] Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.300321 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.300395 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.300506 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crgdp\" (UniqueName: \"kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.402957 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.403060 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.403145 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crgdp\" (UniqueName: \"kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.407571 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.408343 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.418225 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crgdp\" (UniqueName: \"kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dskz4\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:07 crc kubenswrapper[4972]: I0228 11:07:07.503999 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:08 crc kubenswrapper[4972]: I0228 11:07:08.064371 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4"] Feb 28 11:07:08 crc kubenswrapper[4972]: I0228 11:07:08.103118 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" event={"ID":"a19c5936-e88b-4336-814d-8b0053fe2247","Type":"ContainerStarted","Data":"b6f2968a61879279b931ea0cdd7e50220d1601c36163643bcbc14a6e2ea93610"} Feb 28 11:07:09 crc kubenswrapper[4972]: I0228 11:07:09.115237 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" event={"ID":"a19c5936-e88b-4336-814d-8b0053fe2247","Type":"ContainerStarted","Data":"5a738c9e19e5cde2f23710ee5520ef4665d7752d7f752a3b597015ca70eb3323"} Feb 28 11:07:09 crc kubenswrapper[4972]: I0228 11:07:09.131935 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" podStartSLOduration=1.7231414470000002 podStartE2EDuration="2.13191342s" podCreationTimestamp="2026-02-28 11:07:07 +0000 UTC" firstStartedPulling="2026-02-28 11:07:08.071317566 +0000 UTC m=+1884.983325304" lastFinishedPulling="2026-02-28 11:07:08.480089509 +0000 UTC m=+1885.392097277" observedRunningTime="2026-02-28 11:07:09.130762467 +0000 UTC m=+1886.042770215" watchObservedRunningTime="2026-02-28 11:07:09.13191342 +0000 UTC m=+1886.043921158" Feb 28 11:07:16 crc kubenswrapper[4972]: I0228 11:07:16.789005 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:07:16 crc kubenswrapper[4972]: E0228 11:07:16.789832 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:07:17 crc kubenswrapper[4972]: I0228 11:07:17.215169 4972 generic.go:334] "Generic (PLEG): container finished" podID="a19c5936-e88b-4336-814d-8b0053fe2247" containerID="5a738c9e19e5cde2f23710ee5520ef4665d7752d7f752a3b597015ca70eb3323" exitCode=0 Feb 28 11:07:17 crc kubenswrapper[4972]: I0228 11:07:17.215223 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" event={"ID":"a19c5936-e88b-4336-814d-8b0053fe2247","Type":"ContainerDied","Data":"5a738c9e19e5cde2f23710ee5520ef4665d7752d7f752a3b597015ca70eb3323"} Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.671365 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.840196 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory\") pod \"a19c5936-e88b-4336-814d-8b0053fe2247\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.840659 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crgdp\" (UniqueName: \"kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp\") pod \"a19c5936-e88b-4336-814d-8b0053fe2247\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.840706 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam\") pod \"a19c5936-e88b-4336-814d-8b0053fe2247\" (UID: \"a19c5936-e88b-4336-814d-8b0053fe2247\") " Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.849031 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp" (OuterVolumeSpecName: "kube-api-access-crgdp") pod "a19c5936-e88b-4336-814d-8b0053fe2247" (UID: "a19c5936-e88b-4336-814d-8b0053fe2247"). InnerVolumeSpecName "kube-api-access-crgdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.876741 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory" (OuterVolumeSpecName: "inventory") pod "a19c5936-e88b-4336-814d-8b0053fe2247" (UID: "a19c5936-e88b-4336-814d-8b0053fe2247"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.894957 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a19c5936-e88b-4336-814d-8b0053fe2247" (UID: "a19c5936-e88b-4336-814d-8b0053fe2247"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.943078 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crgdp\" (UniqueName: \"kubernetes.io/projected/a19c5936-e88b-4336-814d-8b0053fe2247-kube-api-access-crgdp\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.943129 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:18 crc kubenswrapper[4972]: I0228 11:07:18.943146 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a19c5936-e88b-4336-814d-8b0053fe2247-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.236971 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" event={"ID":"a19c5936-e88b-4336-814d-8b0053fe2247","Type":"ContainerDied","Data":"b6f2968a61879279b931ea0cdd7e50220d1601c36163643bcbc14a6e2ea93610"} Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.237019 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6f2968a61879279b931ea0cdd7e50220d1601c36163643bcbc14a6e2ea93610" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.237036 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dskz4" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.338049 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w"] Feb 28 11:07:19 crc kubenswrapper[4972]: E0228 11:07:19.338744 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19c5936-e88b-4336-814d-8b0053fe2247" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.338772 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19c5936-e88b-4336-814d-8b0053fe2247" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.339001 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a19c5936-e88b-4336-814d-8b0053fe2247" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.340013 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.342875 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.343484 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.343610 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.343711 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.353771 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w"] Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.456833 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.456943 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.457002 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p24dz\" (UniqueName: \"kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.558597 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.558722 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.558769 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p24dz\" (UniqueName: \"kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.563428 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.563953 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.577203 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p24dz\" (UniqueName: \"kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:19 crc kubenswrapper[4972]: I0228 11:07:19.665163 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:20 crc kubenswrapper[4972]: I0228 11:07:20.236271 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w"] Feb 28 11:07:21 crc kubenswrapper[4972]: I0228 11:07:21.258244 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" event={"ID":"dea3597c-d878-41dd-8952-dda5b8db255e","Type":"ContainerStarted","Data":"cf8855a444df0b4798d50f14d2b32d635aeae114dad58c9d04527b0d3a7773f7"} Feb 28 11:07:21 crc kubenswrapper[4972]: I0228 11:07:21.258879 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" event={"ID":"dea3597c-d878-41dd-8952-dda5b8db255e","Type":"ContainerStarted","Data":"a3cccfacf8c25ebea45410ccac54e585e7d2f84a823a4ae1b4aa0d05670ceda9"} Feb 28 11:07:21 crc kubenswrapper[4972]: I0228 11:07:21.292948 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" podStartSLOduration=1.873036224 podStartE2EDuration="2.292923894s" podCreationTimestamp="2026-02-28 11:07:19 +0000 UTC" firstStartedPulling="2026-02-28 11:07:20.243482307 +0000 UTC m=+1897.155490065" lastFinishedPulling="2026-02-28 11:07:20.663369987 +0000 UTC m=+1897.575377735" observedRunningTime="2026-02-28 11:07:21.286652862 +0000 UTC m=+1898.198660600" watchObservedRunningTime="2026-02-28 11:07:21.292923894 +0000 UTC m=+1898.204931622" Feb 28 11:07:26 crc kubenswrapper[4972]: I0228 11:07:26.050130 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8v846"] Feb 28 11:07:26 crc kubenswrapper[4972]: I0228 11:07:26.067937 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8v846"] Feb 28 11:07:27 crc kubenswrapper[4972]: I0228 11:07:27.811016 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6537ac03-6028-4248-8e38-93ade43d502f" path="/var/lib/kubelet/pods/6537ac03-6028-4248-8e38-93ade43d502f/volumes" Feb 28 11:07:30 crc kubenswrapper[4972]: I0228 11:07:30.344862 4972 generic.go:334] "Generic (PLEG): container finished" podID="dea3597c-d878-41dd-8952-dda5b8db255e" containerID="cf8855a444df0b4798d50f14d2b32d635aeae114dad58c9d04527b0d3a7773f7" exitCode=0 Feb 28 11:07:30 crc kubenswrapper[4972]: I0228 11:07:30.344906 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" event={"ID":"dea3597c-d878-41dd-8952-dda5b8db255e","Type":"ContainerDied","Data":"cf8855a444df0b4798d50f14d2b32d635aeae114dad58c9d04527b0d3a7773f7"} Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.790646 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:07:31 crc kubenswrapper[4972]: E0228 11:07:31.791619 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.807903 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.951846 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p24dz\" (UniqueName: \"kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz\") pod \"dea3597c-d878-41dd-8952-dda5b8db255e\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.952129 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory\") pod \"dea3597c-d878-41dd-8952-dda5b8db255e\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.952184 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam\") pod \"dea3597c-d878-41dd-8952-dda5b8db255e\" (UID: \"dea3597c-d878-41dd-8952-dda5b8db255e\") " Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.961164 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz" (OuterVolumeSpecName: "kube-api-access-p24dz") pod "dea3597c-d878-41dd-8952-dda5b8db255e" (UID: "dea3597c-d878-41dd-8952-dda5b8db255e"). InnerVolumeSpecName "kube-api-access-p24dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.981987 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dea3597c-d878-41dd-8952-dda5b8db255e" (UID: "dea3597c-d878-41dd-8952-dda5b8db255e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:31 crc kubenswrapper[4972]: I0228 11:07:31.985917 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory" (OuterVolumeSpecName: "inventory") pod "dea3597c-d878-41dd-8952-dda5b8db255e" (UID: "dea3597c-d878-41dd-8952-dda5b8db255e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.060363 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p24dz\" (UniqueName: \"kubernetes.io/projected/dea3597c-d878-41dd-8952-dda5b8db255e-kube-api-access-p24dz\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.060404 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.060415 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dea3597c-d878-41dd-8952-dda5b8db255e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.367717 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" event={"ID":"dea3597c-d878-41dd-8952-dda5b8db255e","Type":"ContainerDied","Data":"a3cccfacf8c25ebea45410ccac54e585e7d2f84a823a4ae1b4aa0d05670ceda9"} Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.367778 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cccfacf8c25ebea45410ccac54e585e7d2f84a823a4ae1b4aa0d05670ceda9" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.367878 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.468865 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8"] Feb 28 11:07:32 crc kubenswrapper[4972]: E0228 11:07:32.470914 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea3597c-d878-41dd-8952-dda5b8db255e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.470943 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea3597c-d878-41dd-8952-dda5b8db255e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.471189 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea3597c-d878-41dd-8952-dda5b8db255e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.472089 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.478402 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.479070 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.479137 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.479703 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.479791 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.480053 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.480327 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.480631 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.495503 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8"] Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673776 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673852 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fg9l\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673880 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673915 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673943 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.673967 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674036 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674063 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674093 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674116 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674191 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674834 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674864 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.674896 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.775800 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.775918 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.775995 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776178 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776245 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776333 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776617 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776662 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776728 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776778 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776820 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776894 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776930 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fg9l\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.776978 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785138 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785311 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785563 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785747 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785821 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.785972 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.786199 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.786867 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.788478 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.788779 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.794016 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.795120 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.797612 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.803994 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fg9l\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-grxc8\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:32 crc kubenswrapper[4972]: I0228 11:07:32.806835 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:07:33 crc kubenswrapper[4972]: I0228 11:07:33.458315 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8"] Feb 28 11:07:34 crc kubenswrapper[4972]: I0228 11:07:34.397994 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" event={"ID":"585066f8-afbf-42ac-8163-895ef6b76ec5","Type":"ContainerStarted","Data":"d16ada32e3c8f8f24c7b2ff394699cfdbf002c726f6f4b777c5aaa9babfed6fd"} Feb 28 11:07:34 crc kubenswrapper[4972]: I0228 11:07:34.398576 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" event={"ID":"585066f8-afbf-42ac-8163-895ef6b76ec5","Type":"ContainerStarted","Data":"9a17e4b73e82beb180b2ae2d0052963e53130c383c03cbbcd3faab48ac7137ea"} Feb 28 11:07:34 crc kubenswrapper[4972]: I0228 11:07:34.432023 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" podStartSLOduration=1.956072565 podStartE2EDuration="2.432005149s" podCreationTimestamp="2026-02-28 11:07:32 +0000 UTC" firstStartedPulling="2026-02-28 11:07:33.467346379 +0000 UTC m=+1910.379354157" lastFinishedPulling="2026-02-28 11:07:33.943279003 +0000 UTC m=+1910.855286741" observedRunningTime="2026-02-28 11:07:34.420912463 +0000 UTC m=+1911.332920191" watchObservedRunningTime="2026-02-28 11:07:34.432005149 +0000 UTC m=+1911.344012887" Feb 28 11:07:43 crc kubenswrapper[4972]: I0228 11:07:43.798739 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:07:43 crc kubenswrapper[4972]: E0228 11:07:43.800063 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:07:57 crc kubenswrapper[4972]: I0228 11:07:57.790505 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:07:57 crc kubenswrapper[4972]: E0228 11:07:57.791884 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.189910 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537948-tpw5b"] Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.191828 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.194715 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.195404 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.197247 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.207439 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537948-tpw5b"] Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.273340 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzt75\" (UniqueName: \"kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75\") pod \"auto-csr-approver-29537948-tpw5b\" (UID: \"00ab2894-e898-444d-9586-40e9f2333332\") " pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.375610 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzt75\" (UniqueName: \"kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75\") pod \"auto-csr-approver-29537948-tpw5b\" (UID: \"00ab2894-e898-444d-9586-40e9f2333332\") " pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.398886 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzt75\" (UniqueName: \"kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75\") pod \"auto-csr-approver-29537948-tpw5b\" (UID: \"00ab2894-e898-444d-9586-40e9f2333332\") " pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.511748 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:00 crc kubenswrapper[4972]: I0228 11:08:00.977146 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537948-tpw5b"] Feb 28 11:08:01 crc kubenswrapper[4972]: I0228 11:08:01.708876 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" event={"ID":"00ab2894-e898-444d-9586-40e9f2333332","Type":"ContainerStarted","Data":"94565bcbe1d5ef89fc727eb2c41132a4a7d0069aad40167176ad307847c3b896"} Feb 28 11:08:02 crc kubenswrapper[4972]: I0228 11:08:02.720755 4972 generic.go:334] "Generic (PLEG): container finished" podID="00ab2894-e898-444d-9586-40e9f2333332" containerID="a0adbf6be0ff881a57e0c3d071546747690b88ce1f3c51f98220d088cbbf18a4" exitCode=0 Feb 28 11:08:02 crc kubenswrapper[4972]: I0228 11:08:02.720818 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" event={"ID":"00ab2894-e898-444d-9586-40e9f2333332","Type":"ContainerDied","Data":"a0adbf6be0ff881a57e0c3d071546747690b88ce1f3c51f98220d088cbbf18a4"} Feb 28 11:08:03 crc kubenswrapper[4972]: I0228 11:08:03.166681 4972 scope.go:117] "RemoveContainer" containerID="8dcd5a9c060a7077f0e6e00d7a468ccc5176b9666930904d50d93fc122d0a42d" Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.046442 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.158314 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzt75\" (UniqueName: \"kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75\") pod \"00ab2894-e898-444d-9586-40e9f2333332\" (UID: \"00ab2894-e898-444d-9586-40e9f2333332\") " Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.164584 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75" (OuterVolumeSpecName: "kube-api-access-fzt75") pod "00ab2894-e898-444d-9586-40e9f2333332" (UID: "00ab2894-e898-444d-9586-40e9f2333332"). InnerVolumeSpecName "kube-api-access-fzt75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.260520 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzt75\" (UniqueName: \"kubernetes.io/projected/00ab2894-e898-444d-9586-40e9f2333332-kube-api-access-fzt75\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.741913 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" event={"ID":"00ab2894-e898-444d-9586-40e9f2333332","Type":"ContainerDied","Data":"94565bcbe1d5ef89fc727eb2c41132a4a7d0069aad40167176ad307847c3b896"} Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.741955 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94565bcbe1d5ef89fc727eb2c41132a4a7d0069aad40167176ad307847c3b896" Feb 28 11:08:04 crc kubenswrapper[4972]: I0228 11:08:04.742000 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537948-tpw5b" Feb 28 11:08:05 crc kubenswrapper[4972]: I0228 11:08:05.146850 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537942-sqfjd"] Feb 28 11:08:05 crc kubenswrapper[4972]: I0228 11:08:05.161096 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537942-sqfjd"] Feb 28 11:08:05 crc kubenswrapper[4972]: I0228 11:08:05.799931 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9583bd17-c0e0-405c-8ea0-f25420fae38f" path="/var/lib/kubelet/pods/9583bd17-c0e0-405c-8ea0-f25420fae38f/volumes" Feb 28 11:08:10 crc kubenswrapper[4972]: I0228 11:08:10.790904 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:08:10 crc kubenswrapper[4972]: E0228 11:08:10.791843 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:08:13 crc kubenswrapper[4972]: I0228 11:08:13.854535 4972 generic.go:334] "Generic (PLEG): container finished" podID="585066f8-afbf-42ac-8163-895ef6b76ec5" containerID="d16ada32e3c8f8f24c7b2ff394699cfdbf002c726f6f4b777c5aaa9babfed6fd" exitCode=0 Feb 28 11:08:13 crc kubenswrapper[4972]: I0228 11:08:13.854602 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" event={"ID":"585066f8-afbf-42ac-8163-895ef6b76ec5","Type":"ContainerDied","Data":"d16ada32e3c8f8f24c7b2ff394699cfdbf002c726f6f4b777c5aaa9babfed6fd"} Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.346669 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.434639 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.434711 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.434803 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.434895 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.434977 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435090 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fg9l\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435179 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435272 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435318 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435382 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435485 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435526 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435586 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.435665 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"585066f8-afbf-42ac-8163-895ef6b76ec5\" (UID: \"585066f8-afbf-42ac-8163-895ef6b76ec5\") " Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.446284 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.446349 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.446452 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.447091 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l" (OuterVolumeSpecName: "kube-api-access-9fg9l") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "kube-api-access-9fg9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.447209 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.447447 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.447511 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.447604 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.448125 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.449325 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.455510 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.468586 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.472729 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory" (OuterVolumeSpecName: "inventory") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.473524 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "585066f8-afbf-42ac-8163-895ef6b76ec5" (UID: "585066f8-afbf-42ac-8163-895ef6b76ec5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539022 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539278 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539379 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539448 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fg9l\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-kube-api-access-9fg9l\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539541 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539614 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539679 4972 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539739 4972 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539815 4972 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539892 4972 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.539960 4972 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.540014 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/585066f8-afbf-42ac-8163-895ef6b76ec5-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.540068 4972 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.540125 4972 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585066f8-afbf-42ac-8163-895ef6b76ec5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.875554 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" event={"ID":"585066f8-afbf-42ac-8163-895ef6b76ec5","Type":"ContainerDied","Data":"9a17e4b73e82beb180b2ae2d0052963e53130c383c03cbbcd3faab48ac7137ea"} Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.875614 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a17e4b73e82beb180b2ae2d0052963e53130c383c03cbbcd3faab48ac7137ea" Feb 28 11:08:15 crc kubenswrapper[4972]: I0228 11:08:15.875629 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-grxc8" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.050960 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j"] Feb 28 11:08:16 crc kubenswrapper[4972]: E0228 11:08:16.051325 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="585066f8-afbf-42ac-8163-895ef6b76ec5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.051344 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="585066f8-afbf-42ac-8163-895ef6b76ec5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 28 11:08:16 crc kubenswrapper[4972]: E0228 11:08:16.051364 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ab2894-e898-444d-9586-40e9f2333332" containerName="oc" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.051371 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ab2894-e898-444d-9586-40e9f2333332" containerName="oc" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.051584 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ab2894-e898-444d-9586-40e9f2333332" containerName="oc" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.051595 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="585066f8-afbf-42ac-8163-895ef6b76ec5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.052199 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.057140 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.057152 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.057140 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.057513 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.067049 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.068050 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j"] Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.158372 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.158485 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.159625 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.159755 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.160020 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5szq6\" (UniqueName: \"kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.262105 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5szq6\" (UniqueName: \"kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.262436 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.262556 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.262585 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.262630 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.263640 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.267285 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.269873 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.270929 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.278021 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5szq6\" (UniqueName: \"kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k9z8j\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.368494 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:08:16 crc kubenswrapper[4972]: I0228 11:08:16.922939 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j"] Feb 28 11:08:17 crc kubenswrapper[4972]: I0228 11:08:17.895445 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" event={"ID":"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b","Type":"ContainerStarted","Data":"a299c9c04a25d7bcea5c1980ac915aa1b8a3800854f63f118f4fdfd558645664"} Feb 28 11:08:17 crc kubenswrapper[4972]: I0228 11:08:17.895839 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" event={"ID":"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b","Type":"ContainerStarted","Data":"a2eaa7208d45f38e22434da09ccc050279910a420a5becb3df5ac2be8d90cd61"} Feb 28 11:08:17 crc kubenswrapper[4972]: I0228 11:08:17.920851 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" podStartSLOduration=1.267029887 podStartE2EDuration="1.920828802s" podCreationTimestamp="2026-02-28 11:08:16 +0000 UTC" firstStartedPulling="2026-02-28 11:08:16.919416049 +0000 UTC m=+1953.831423797" lastFinishedPulling="2026-02-28 11:08:17.573214974 +0000 UTC m=+1954.485222712" observedRunningTime="2026-02-28 11:08:17.91056487 +0000 UTC m=+1954.822572648" watchObservedRunningTime="2026-02-28 11:08:17.920828802 +0000 UTC m=+1954.832836550" Feb 28 11:08:24 crc kubenswrapper[4972]: I0228 11:08:24.790862 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:08:24 crc kubenswrapper[4972]: E0228 11:08:24.792836 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:08:35 crc kubenswrapper[4972]: I0228 11:08:35.790493 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:08:35 crc kubenswrapper[4972]: E0228 11:08:35.791894 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:08:49 crc kubenswrapper[4972]: I0228 11:08:49.790092 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:08:49 crc kubenswrapper[4972]: E0228 11:08:49.791503 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:09:03 crc kubenswrapper[4972]: I0228 11:09:03.253253 4972 scope.go:117] "RemoveContainer" containerID="d60bc133fdb6b73b61a28f564e0b5a8c27daac1cad1be8152891ca20bcd5d579" Feb 28 11:09:03 crc kubenswrapper[4972]: I0228 11:09:03.303841 4972 scope.go:117] "RemoveContainer" containerID="cef2dd316b1090e84773595d3417aca7dd203800e2ef4c5c89d9285b34c88807" Feb 28 11:09:03 crc kubenswrapper[4972]: I0228 11:09:03.334859 4972 scope.go:117] "RemoveContainer" containerID="2a6b785cf626e1e877a3e16c66489a63ab22056ba2ab3b4c39db8a6f16024f57" Feb 28 11:09:03 crc kubenswrapper[4972]: I0228 11:09:03.434069 4972 scope.go:117] "RemoveContainer" containerID="1ee022c0cf96815f93a5a775cc7465f15b62727af031bd036979c7956af379fc" Feb 28 11:09:03 crc kubenswrapper[4972]: I0228 11:09:03.795664 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:09:03 crc kubenswrapper[4972]: E0228 11:09:03.796017 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:09:18 crc kubenswrapper[4972]: I0228 11:09:18.789754 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:09:18 crc kubenswrapper[4972]: E0228 11:09:18.790633 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:09:25 crc kubenswrapper[4972]: I0228 11:09:25.649370 4972 generic.go:334] "Generic (PLEG): container finished" podID="d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" containerID="a299c9c04a25d7bcea5c1980ac915aa1b8a3800854f63f118f4fdfd558645664" exitCode=0 Feb 28 11:09:25 crc kubenswrapper[4972]: I0228 11:09:25.649499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" event={"ID":"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b","Type":"ContainerDied","Data":"a299c9c04a25d7bcea5c1980ac915aa1b8a3800854f63f118f4fdfd558645664"} Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.121752 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.218548 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle\") pod \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.218636 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam\") pod \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.218826 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0\") pod \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.218897 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory\") pod \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.218979 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5szq6\" (UniqueName: \"kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6\") pod \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\" (UID: \"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b\") " Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.229418 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" (UID: "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.229640 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6" (OuterVolumeSpecName: "kube-api-access-5szq6") pod "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" (UID: "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b"). InnerVolumeSpecName "kube-api-access-5szq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.245592 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" (UID: "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.250447 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" (UID: "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.266109 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory" (OuterVolumeSpecName: "inventory") pod "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" (UID: "d2cb5f2e-f9b1-41fb-8a63-03bebe97019b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.321623 4972 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.321659 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.321672 4972 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.321681 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.321692 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5szq6\" (UniqueName: \"kubernetes.io/projected/d2cb5f2e-f9b1-41fb-8a63-03bebe97019b-kube-api-access-5szq6\") on node \"crc\" DevicePath \"\"" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.695433 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" event={"ID":"d2cb5f2e-f9b1-41fb-8a63-03bebe97019b","Type":"ContainerDied","Data":"a2eaa7208d45f38e22434da09ccc050279910a420a5becb3df5ac2be8d90cd61"} Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.695891 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2eaa7208d45f38e22434da09ccc050279910a420a5becb3df5ac2be8d90cd61" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.695695 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k9z8j" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.823310 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g"] Feb 28 11:09:27 crc kubenswrapper[4972]: E0228 11:09:27.823825 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.823848 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.824107 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cb5f2e-f9b1-41fb-8a63-03bebe97019b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.825007 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.828930 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.831531 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.831528 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.833969 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.834247 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.834337 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.836210 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g"] Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.934701 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.934864 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.934953 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.935047 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.935095 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkmnp\" (UniqueName: \"kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:27 crc kubenswrapper[4972]: I0228 11:09:27.935160 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037574 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037664 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037710 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037788 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkmnp\" (UniqueName: \"kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.037825 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.043251 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.043304 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.044006 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.044818 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.045072 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.062431 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkmnp\" (UniqueName: \"kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.148085 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:09:28 crc kubenswrapper[4972]: I0228 11:09:28.781425 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g"] Feb 28 11:09:29 crc kubenswrapper[4972]: I0228 11:09:29.741248 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" event={"ID":"ae3ba487-6066-46c6-b9f2-5acc9810d3bf","Type":"ContainerStarted","Data":"1187afd0ec43455f9fdcfcb931975f474d9aee99a611af5d46bbf9e2e63c5f0f"} Feb 28 11:09:29 crc kubenswrapper[4972]: I0228 11:09:29.742158 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" event={"ID":"ae3ba487-6066-46c6-b9f2-5acc9810d3bf","Type":"ContainerStarted","Data":"f44e875215c44202300ff1a6b005cc463cebbb265dc682552ca9466c4d57a970"} Feb 28 11:09:29 crc kubenswrapper[4972]: I0228 11:09:29.780096 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" podStartSLOduration=2.261842116 podStartE2EDuration="2.780063045s" podCreationTimestamp="2026-02-28 11:09:27 +0000 UTC" firstStartedPulling="2026-02-28 11:09:28.77679979 +0000 UTC m=+2025.688807528" lastFinishedPulling="2026-02-28 11:09:29.295020709 +0000 UTC m=+2026.207028457" observedRunningTime="2026-02-28 11:09:29.766587523 +0000 UTC m=+2026.678595281" watchObservedRunningTime="2026-02-28 11:09:29.780063045 +0000 UTC m=+2026.692070803" Feb 28 11:09:31 crc kubenswrapper[4972]: I0228 11:09:31.790018 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:09:31 crc kubenswrapper[4972]: E0228 11:09:31.791219 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:09:44 crc kubenswrapper[4972]: I0228 11:09:44.790724 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:09:44 crc kubenswrapper[4972]: E0228 11:09:44.792374 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:09:56 crc kubenswrapper[4972]: I0228 11:09:56.790416 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:09:56 crc kubenswrapper[4972]: E0228 11:09:56.791547 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.156185 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537950-r9rpl"] Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.158935 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.164827 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537950-r9rpl"] Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.187512 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjf8z\" (UniqueName: \"kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z\") pod \"auto-csr-approver-29537950-r9rpl\" (UID: \"37f24d57-f1a3-4785-b4be-54d73bbde481\") " pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.188629 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.188636 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.188687 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.289760 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjf8z\" (UniqueName: \"kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z\") pod \"auto-csr-approver-29537950-r9rpl\" (UID: \"37f24d57-f1a3-4785-b4be-54d73bbde481\") " pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.310782 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjf8z\" (UniqueName: \"kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z\") pod \"auto-csr-approver-29537950-r9rpl\" (UID: \"37f24d57-f1a3-4785-b4be-54d73bbde481\") " pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.500913 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:00 crc kubenswrapper[4972]: W0228 11:10:00.970451 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37f24d57_f1a3_4785_b4be_54d73bbde481.slice/crio-f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039 WatchSource:0}: Error finding container f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039: Status 404 returned error can't find the container with id f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039 Feb 28 11:10:00 crc kubenswrapper[4972]: I0228 11:10:00.975673 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537950-r9rpl"] Feb 28 11:10:01 crc kubenswrapper[4972]: I0228 11:10:01.098218 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" event={"ID":"37f24d57-f1a3-4785-b4be-54d73bbde481","Type":"ContainerStarted","Data":"f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039"} Feb 28 11:10:03 crc kubenswrapper[4972]: I0228 11:10:03.124150 4972 generic.go:334] "Generic (PLEG): container finished" podID="37f24d57-f1a3-4785-b4be-54d73bbde481" containerID="18e9497d2508fbbf53b5cef655b74f241514b6313ef72e2a3eb39a7962001565" exitCode=0 Feb 28 11:10:03 crc kubenswrapper[4972]: I0228 11:10:03.124226 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" event={"ID":"37f24d57-f1a3-4785-b4be-54d73bbde481","Type":"ContainerDied","Data":"18e9497d2508fbbf53b5cef655b74f241514b6313ef72e2a3eb39a7962001565"} Feb 28 11:10:04 crc kubenswrapper[4972]: I0228 11:10:04.479578 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:04 crc kubenswrapper[4972]: I0228 11:10:04.594780 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjf8z\" (UniqueName: \"kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z\") pod \"37f24d57-f1a3-4785-b4be-54d73bbde481\" (UID: \"37f24d57-f1a3-4785-b4be-54d73bbde481\") " Feb 28 11:10:04 crc kubenswrapper[4972]: I0228 11:10:04.602992 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z" (OuterVolumeSpecName: "kube-api-access-pjf8z") pod "37f24d57-f1a3-4785-b4be-54d73bbde481" (UID: "37f24d57-f1a3-4785-b4be-54d73bbde481"). InnerVolumeSpecName "kube-api-access-pjf8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:10:04 crc kubenswrapper[4972]: I0228 11:10:04.698594 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjf8z\" (UniqueName: \"kubernetes.io/projected/37f24d57-f1a3-4785-b4be-54d73bbde481-kube-api-access-pjf8z\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.154284 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" event={"ID":"37f24d57-f1a3-4785-b4be-54d73bbde481","Type":"ContainerDied","Data":"f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039"} Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.154810 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4da6f4ed7762de7ac8d5c62834d269958ec6953d5ed1c121e3fd86208a42039" Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.154358 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537950-r9rpl" Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.568315 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537944-s7lj4"] Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.583042 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537944-s7lj4"] Feb 28 11:10:05 crc kubenswrapper[4972]: I0228 11:10:05.799395 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3864232-cd37-48db-a3fc-8e80a05f2312" path="/var/lib/kubelet/pods/e3864232-cd37-48db-a3fc-8e80a05f2312/volumes" Feb 28 11:10:08 crc kubenswrapper[4972]: I0228 11:10:08.790639 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:10:08 crc kubenswrapper[4972]: E0228 11:10:08.791383 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:10:19 crc kubenswrapper[4972]: I0228 11:10:19.302746 4972 generic.go:334] "Generic (PLEG): container finished" podID="ae3ba487-6066-46c6-b9f2-5acc9810d3bf" containerID="1187afd0ec43455f9fdcfcb931975f474d9aee99a611af5d46bbf9e2e63c5f0f" exitCode=0 Feb 28 11:10:19 crc kubenswrapper[4972]: I0228 11:10:19.302827 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" event={"ID":"ae3ba487-6066-46c6-b9f2-5acc9810d3bf","Type":"ContainerDied","Data":"1187afd0ec43455f9fdcfcb931975f474d9aee99a611af5d46bbf9e2e63c5f0f"} Feb 28 11:10:19 crc kubenswrapper[4972]: I0228 11:10:19.789858 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:10:19 crc kubenswrapper[4972]: E0228 11:10:19.790349 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.867175 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975581 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975624 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkmnp\" (UniqueName: \"kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975649 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975692 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975721 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.975761 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\" (UID: \"ae3ba487-6066-46c6-b9f2-5acc9810d3bf\") " Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.990055 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:10:20 crc kubenswrapper[4972]: I0228 11:10:20.990076 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp" (OuterVolumeSpecName: "kube-api-access-nkmnp") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "kube-api-access-nkmnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.004345 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.005600 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory" (OuterVolumeSpecName: "inventory") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.006276 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.026177 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ae3ba487-6066-46c6-b9f2-5acc9810d3bf" (UID: "ae3ba487-6066-46c6-b9f2-5acc9810d3bf"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078113 4972 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078395 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078479 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkmnp\" (UniqueName: \"kubernetes.io/projected/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-kube-api-access-nkmnp\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078544 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078604 4972 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.078659 4972 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ae3ba487-6066-46c6-b9f2-5acc9810d3bf-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.332174 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" event={"ID":"ae3ba487-6066-46c6-b9f2-5acc9810d3bf","Type":"ContainerDied","Data":"f44e875215c44202300ff1a6b005cc463cebbb265dc682552ca9466c4d57a970"} Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.332840 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f44e875215c44202300ff1a6b005cc463cebbb265dc682552ca9466c4d57a970" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.332291 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.445226 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz"] Feb 28 11:10:21 crc kubenswrapper[4972]: E0228 11:10:21.445756 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3ba487-6066-46c6-b9f2-5acc9810d3bf" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.445780 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3ba487-6066-46c6-b9f2-5acc9810d3bf" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 28 11:10:21 crc kubenswrapper[4972]: E0228 11:10:21.445822 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f24d57-f1a3-4785-b4be-54d73bbde481" containerName="oc" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.445831 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f24d57-f1a3-4785-b4be-54d73bbde481" containerName="oc" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.446093 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f24d57-f1a3-4785-b4be-54d73bbde481" containerName="oc" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.446128 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3ba487-6066-46c6-b9f2-5acc9810d3bf" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.447016 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.448991 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.455222 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz"] Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.467014 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.467774 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.467922 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.469127 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.492348 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.492562 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.492599 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.492623 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824zw\" (UniqueName: \"kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.492728 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.594214 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.594340 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.594363 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.594381 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824zw\" (UniqueName: \"kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.594409 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.598101 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.598101 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.598808 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.598921 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.613742 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824zw\" (UniqueName: \"kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:21 crc kubenswrapper[4972]: I0228 11:10:21.781751 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:10:22 crc kubenswrapper[4972]: I0228 11:10:22.327229 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz"] Feb 28 11:10:22 crc kubenswrapper[4972]: I0228 11:10:22.330276 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:10:22 crc kubenswrapper[4972]: I0228 11:10:22.344577 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" event={"ID":"1a18fc65-40ad-474f-b28d-76e3d12cb6df","Type":"ContainerStarted","Data":"0edf7eec669c9e29c918c5d279e0e2c1a0519849ab084a04af4ee7e8dacd511a"} Feb 28 11:10:23 crc kubenswrapper[4972]: I0228 11:10:23.355562 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" event={"ID":"1a18fc65-40ad-474f-b28d-76e3d12cb6df","Type":"ContainerStarted","Data":"14fb18454c30a80416435bb8285f19d8407334dc50426c792ab856ee39fad4ac"} Feb 28 11:10:23 crc kubenswrapper[4972]: I0228 11:10:23.386787 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" podStartSLOduration=2.005976093 podStartE2EDuration="2.386769776s" podCreationTimestamp="2026-02-28 11:10:21 +0000 UTC" firstStartedPulling="2026-02-28 11:10:22.329953875 +0000 UTC m=+2079.241961613" lastFinishedPulling="2026-02-28 11:10:22.710747558 +0000 UTC m=+2079.622755296" observedRunningTime="2026-02-28 11:10:23.3751511 +0000 UTC m=+2080.287158838" watchObservedRunningTime="2026-02-28 11:10:23.386769776 +0000 UTC m=+2080.298777514" Feb 28 11:10:32 crc kubenswrapper[4972]: I0228 11:10:32.790012 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:10:33 crc kubenswrapper[4972]: I0228 11:10:33.467075 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c"} Feb 28 11:11:03 crc kubenswrapper[4972]: I0228 11:11:03.522543 4972 scope.go:117] "RemoveContainer" containerID="a196592e1e7d7382dad1f7d31cd0377ae1ddc05bed19728be064208108262785" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.639869 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.661182 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.661751 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.685730 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5xt4\" (UniqueName: \"kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.685854 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.685973 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.788187 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5xt4\" (UniqueName: \"kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.788268 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.788355 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.788894 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.788897 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.811801 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5xt4\" (UniqueName: \"kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4\") pod \"redhat-operators-fl2sj\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:26 crc kubenswrapper[4972]: I0228 11:11:26.999278 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:27 crc kubenswrapper[4972]: I0228 11:11:27.535802 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:27 crc kubenswrapper[4972]: W0228 11:11:27.540247 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a02aec2_d47e_406d_92e8_173fe5bd7462.slice/crio-f7bb9c32aee313444c3e3433ae644ae77d0b9ca69e759ab44c3a5e3a852abbe8 WatchSource:0}: Error finding container f7bb9c32aee313444c3e3433ae644ae77d0b9ca69e759ab44c3a5e3a852abbe8: Status 404 returned error can't find the container with id f7bb9c32aee313444c3e3433ae644ae77d0b9ca69e759ab44c3a5e3a852abbe8 Feb 28 11:11:28 crc kubenswrapper[4972]: I0228 11:11:28.051046 4972 generic.go:334] "Generic (PLEG): container finished" podID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerID="770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326" exitCode=0 Feb 28 11:11:28 crc kubenswrapper[4972]: I0228 11:11:28.051128 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerDied","Data":"770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326"} Feb 28 11:11:28 crc kubenswrapper[4972]: I0228 11:11:28.051447 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerStarted","Data":"f7bb9c32aee313444c3e3433ae644ae77d0b9ca69e759ab44c3a5e3a852abbe8"} Feb 28 11:11:29 crc kubenswrapper[4972]: I0228 11:11:29.064124 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerStarted","Data":"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57"} Feb 28 11:11:30 crc kubenswrapper[4972]: I0228 11:11:30.074350 4972 generic.go:334] "Generic (PLEG): container finished" podID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerID="4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57" exitCode=0 Feb 28 11:11:30 crc kubenswrapper[4972]: I0228 11:11:30.074411 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerDied","Data":"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57"} Feb 28 11:11:31 crc kubenswrapper[4972]: I0228 11:11:31.088200 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerStarted","Data":"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1"} Feb 28 11:11:31 crc kubenswrapper[4972]: I0228 11:11:31.135431 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fl2sj" podStartSLOduration=2.712494259 podStartE2EDuration="5.135401318s" podCreationTimestamp="2026-02-28 11:11:26 +0000 UTC" firstStartedPulling="2026-02-28 11:11:28.053432284 +0000 UTC m=+2144.965440022" lastFinishedPulling="2026-02-28 11:11:30.476339333 +0000 UTC m=+2147.388347081" observedRunningTime="2026-02-28 11:11:31.123839534 +0000 UTC m=+2148.035847302" watchObservedRunningTime="2026-02-28 11:11:31.135401318 +0000 UTC m=+2148.047409086" Feb 28 11:11:37 crc kubenswrapper[4972]: I0228 11:11:37.000428 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:37 crc kubenswrapper[4972]: I0228 11:11:37.000957 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:37 crc kubenswrapper[4972]: I0228 11:11:37.060317 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:37 crc kubenswrapper[4972]: I0228 11:11:37.194998 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:37 crc kubenswrapper[4972]: I0228 11:11:37.609117 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.162959 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fl2sj" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="registry-server" containerID="cri-o://a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1" gracePeriod=2 Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.674172 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.864322 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5xt4\" (UniqueName: \"kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4\") pod \"3a02aec2-d47e-406d-92e8-173fe5bd7462\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.864397 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities\") pod \"3a02aec2-d47e-406d-92e8-173fe5bd7462\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.864556 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content\") pod \"3a02aec2-d47e-406d-92e8-173fe5bd7462\" (UID: \"3a02aec2-d47e-406d-92e8-173fe5bd7462\") " Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.865611 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities" (OuterVolumeSpecName: "utilities") pod "3a02aec2-d47e-406d-92e8-173fe5bd7462" (UID: "3a02aec2-d47e-406d-92e8-173fe5bd7462"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.870003 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4" (OuterVolumeSpecName: "kube-api-access-g5xt4") pod "3a02aec2-d47e-406d-92e8-173fe5bd7462" (UID: "3a02aec2-d47e-406d-92e8-173fe5bd7462"). InnerVolumeSpecName "kube-api-access-g5xt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.968426 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5xt4\" (UniqueName: \"kubernetes.io/projected/3a02aec2-d47e-406d-92e8-173fe5bd7462-kube-api-access-g5xt4\") on node \"crc\" DevicePath \"\"" Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.968455 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:11:39 crc kubenswrapper[4972]: I0228 11:11:39.970445 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a02aec2-d47e-406d-92e8-173fe5bd7462" (UID: "3a02aec2-d47e-406d-92e8-173fe5bd7462"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.071302 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a02aec2-d47e-406d-92e8-173fe5bd7462-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.176291 4972 generic.go:334] "Generic (PLEG): container finished" podID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerID="a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1" exitCode=0 Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.176371 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerDied","Data":"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1"} Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.177957 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fl2sj" event={"ID":"3a02aec2-d47e-406d-92e8-173fe5bd7462","Type":"ContainerDied","Data":"f7bb9c32aee313444c3e3433ae644ae77d0b9ca69e759ab44c3a5e3a852abbe8"} Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.177993 4972 scope.go:117] "RemoveContainer" containerID="a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.176439 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fl2sj" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.201670 4972 scope.go:117] "RemoveContainer" containerID="4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.213650 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.221882 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fl2sj"] Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.241712 4972 scope.go:117] "RemoveContainer" containerID="770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.267344 4972 scope.go:117] "RemoveContainer" containerID="a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1" Feb 28 11:11:40 crc kubenswrapper[4972]: E0228 11:11:40.267839 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1\": container with ID starting with a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1 not found: ID does not exist" containerID="a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.267869 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1"} err="failed to get container status \"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1\": rpc error: code = NotFound desc = could not find container \"a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1\": container with ID starting with a6a7a3b857ad2feaf877b97913798951e1fe30c88feeca9e9a4e015753be45d1 not found: ID does not exist" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.267893 4972 scope.go:117] "RemoveContainer" containerID="4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57" Feb 28 11:11:40 crc kubenswrapper[4972]: E0228 11:11:40.268303 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57\": container with ID starting with 4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57 not found: ID does not exist" containerID="4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.268335 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57"} err="failed to get container status \"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57\": rpc error: code = NotFound desc = could not find container \"4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57\": container with ID starting with 4aa6c219f00c0c3baf09fae838da7d7e0576655938d651aa84e52a1334208d57 not found: ID does not exist" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.268352 4972 scope.go:117] "RemoveContainer" containerID="770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326" Feb 28 11:11:40 crc kubenswrapper[4972]: E0228 11:11:40.268775 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326\": container with ID starting with 770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326 not found: ID does not exist" containerID="770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326" Feb 28 11:11:40 crc kubenswrapper[4972]: I0228 11:11:40.268818 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326"} err="failed to get container status \"770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326\": rpc error: code = NotFound desc = could not find container \"770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326\": container with ID starting with 770f2bb31dca6d08a013435a10c596d490ccb602be17cc96a898a23f31906326 not found: ID does not exist" Feb 28 11:11:41 crc kubenswrapper[4972]: I0228 11:11:41.801505 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" path="/var/lib/kubelet/pods/3a02aec2-d47e-406d-92e8-173fe5bd7462/volumes" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.149938 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537952-rnxj8"] Feb 28 11:12:00 crc kubenswrapper[4972]: E0228 11:12:00.151016 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="extract-content" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.151039 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="extract-content" Feb 28 11:12:00 crc kubenswrapper[4972]: E0228 11:12:00.151085 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="registry-server" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.151093 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="registry-server" Feb 28 11:12:00 crc kubenswrapper[4972]: E0228 11:12:00.151122 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="extract-utilities" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.151131 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="extract-utilities" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.151386 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a02aec2-d47e-406d-92e8-173fe5bd7462" containerName="registry-server" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.152238 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.154715 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.155648 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.155740 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.192119 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537952-rnxj8"] Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.259546 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbrhm\" (UniqueName: \"kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm\") pod \"auto-csr-approver-29537952-rnxj8\" (UID: \"7cfa36c3-a86e-4fad-b923-e73fbfb34e83\") " pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.362132 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbrhm\" (UniqueName: \"kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm\") pod \"auto-csr-approver-29537952-rnxj8\" (UID: \"7cfa36c3-a86e-4fad-b923-e73fbfb34e83\") " pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.386750 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbrhm\" (UniqueName: \"kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm\") pod \"auto-csr-approver-29537952-rnxj8\" (UID: \"7cfa36c3-a86e-4fad-b923-e73fbfb34e83\") " pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.477168 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:00 crc kubenswrapper[4972]: I0228 11:12:00.784374 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537952-rnxj8"] Feb 28 11:12:01 crc kubenswrapper[4972]: I0228 11:12:01.425170 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" event={"ID":"7cfa36c3-a86e-4fad-b923-e73fbfb34e83","Type":"ContainerStarted","Data":"19803c2f19d4d2b3a6fd9266825146b8de1afafb96f849a43497ffd2598ba0e2"} Feb 28 11:12:02 crc kubenswrapper[4972]: I0228 11:12:02.438300 4972 generic.go:334] "Generic (PLEG): container finished" podID="7cfa36c3-a86e-4fad-b923-e73fbfb34e83" containerID="2a2ba8b8e8245bb9874abfc50b011576b73ab63d1908d2d191e7d5969f9d0b98" exitCode=0 Feb 28 11:12:02 crc kubenswrapper[4972]: I0228 11:12:02.438483 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" event={"ID":"7cfa36c3-a86e-4fad-b923-e73fbfb34e83","Type":"ContainerDied","Data":"2a2ba8b8e8245bb9874abfc50b011576b73ab63d1908d2d191e7d5969f9d0b98"} Feb 28 11:12:03 crc kubenswrapper[4972]: I0228 11:12:03.903702 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.055583 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbrhm\" (UniqueName: \"kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm\") pod \"7cfa36c3-a86e-4fad-b923-e73fbfb34e83\" (UID: \"7cfa36c3-a86e-4fad-b923-e73fbfb34e83\") " Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.064303 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm" (OuterVolumeSpecName: "kube-api-access-rbrhm") pod "7cfa36c3-a86e-4fad-b923-e73fbfb34e83" (UID: "7cfa36c3-a86e-4fad-b923-e73fbfb34e83"). InnerVolumeSpecName "kube-api-access-rbrhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.158768 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbrhm\" (UniqueName: \"kubernetes.io/projected/7cfa36c3-a86e-4fad-b923-e73fbfb34e83-kube-api-access-rbrhm\") on node \"crc\" DevicePath \"\"" Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.459702 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" event={"ID":"7cfa36c3-a86e-4fad-b923-e73fbfb34e83","Type":"ContainerDied","Data":"19803c2f19d4d2b3a6fd9266825146b8de1afafb96f849a43497ffd2598ba0e2"} Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.459744 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19803c2f19d4d2b3a6fd9266825146b8de1afafb96f849a43497ffd2598ba0e2" Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.459809 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537952-rnxj8" Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.981370 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537946-nzd87"] Feb 28 11:12:04 crc kubenswrapper[4972]: I0228 11:12:04.994091 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537946-nzd87"] Feb 28 11:12:05 crc kubenswrapper[4972]: I0228 11:12:05.803733 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6510a03f-c392-4cde-b9be-a0a6405cf2c3" path="/var/lib/kubelet/pods/6510a03f-c392-4cde-b9be-a0a6405cf2c3/volumes" Feb 28 11:12:05 crc kubenswrapper[4972]: I0228 11:12:05.990998 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:05 crc kubenswrapper[4972]: E0228 11:12:05.991725 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfa36c3-a86e-4fad-b923-e73fbfb34e83" containerName="oc" Feb 28 11:12:05 crc kubenswrapper[4972]: I0228 11:12:05.991748 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfa36c3-a86e-4fad-b923-e73fbfb34e83" containerName="oc" Feb 28 11:12:05 crc kubenswrapper[4972]: I0228 11:12:05.992047 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfa36c3-a86e-4fad-b923-e73fbfb34e83" containerName="oc" Feb 28 11:12:05 crc kubenswrapper[4972]: I0228 11:12:05.994913 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.007004 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.108331 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.108777 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.108920 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4pg\" (UniqueName: \"kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.212089 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.212209 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4pg\" (UniqueName: \"kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.212333 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.213018 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.213059 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.253599 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4pg\" (UniqueName: \"kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg\") pod \"certified-operators-ctgz6\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.330198 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:06 crc kubenswrapper[4972]: I0228 11:12:06.744572 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:06 crc kubenswrapper[4972]: W0228 11:12:06.754628 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc70d7dea_801a_4fe3_9b1f_5e37492d3df2.slice/crio-95c431c167c8f6db7f8cd75b5b5a2bd742d8637ec3c1d466ae44da561e31e99e WatchSource:0}: Error finding container 95c431c167c8f6db7f8cd75b5b5a2bd742d8637ec3c1d466ae44da561e31e99e: Status 404 returned error can't find the container with id 95c431c167c8f6db7f8cd75b5b5a2bd742d8637ec3c1d466ae44da561e31e99e Feb 28 11:12:07 crc kubenswrapper[4972]: I0228 11:12:07.497568 4972 generic.go:334] "Generic (PLEG): container finished" podID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerID="c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64" exitCode=0 Feb 28 11:12:07 crc kubenswrapper[4972]: I0228 11:12:07.497691 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerDied","Data":"c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64"} Feb 28 11:12:07 crc kubenswrapper[4972]: I0228 11:12:07.497988 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerStarted","Data":"95c431c167c8f6db7f8cd75b5b5a2bd742d8637ec3c1d466ae44da561e31e99e"} Feb 28 11:12:08 crc kubenswrapper[4972]: I0228 11:12:08.511413 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerStarted","Data":"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729"} Feb 28 11:12:09 crc kubenswrapper[4972]: I0228 11:12:09.530739 4972 generic.go:334] "Generic (PLEG): container finished" podID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerID="1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729" exitCode=0 Feb 28 11:12:09 crc kubenswrapper[4972]: I0228 11:12:09.530843 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerDied","Data":"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729"} Feb 28 11:12:11 crc kubenswrapper[4972]: I0228 11:12:11.554199 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerStarted","Data":"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a"} Feb 28 11:12:11 crc kubenswrapper[4972]: I0228 11:12:11.585325 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ctgz6" podStartSLOduration=4.161136756 podStartE2EDuration="6.585292008s" podCreationTimestamp="2026-02-28 11:12:05 +0000 UTC" firstStartedPulling="2026-02-28 11:12:07.500370409 +0000 UTC m=+2184.412378177" lastFinishedPulling="2026-02-28 11:12:09.924525681 +0000 UTC m=+2186.836533429" observedRunningTime="2026-02-28 11:12:11.580313613 +0000 UTC m=+2188.492321351" watchObservedRunningTime="2026-02-28 11:12:11.585292008 +0000 UTC m=+2188.497299746" Feb 28 11:12:16 crc kubenswrapper[4972]: I0228 11:12:16.330910 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:16 crc kubenswrapper[4972]: I0228 11:12:16.331536 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:16 crc kubenswrapper[4972]: I0228 11:12:16.379234 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:16 crc kubenswrapper[4972]: I0228 11:12:16.674336 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:16 crc kubenswrapper[4972]: I0228 11:12:16.725076 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:18 crc kubenswrapper[4972]: I0228 11:12:18.629948 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ctgz6" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="registry-server" containerID="cri-o://9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a" gracePeriod=2 Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.106964 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.129808 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities\") pod \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.129910 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l4pg\" (UniqueName: \"kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg\") pod \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.129947 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content\") pod \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\" (UID: \"c70d7dea-801a-4fe3-9b1f-5e37492d3df2\") " Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.130688 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities" (OuterVolumeSpecName: "utilities") pod "c70d7dea-801a-4fe3-9b1f-5e37492d3df2" (UID: "c70d7dea-801a-4fe3-9b1f-5e37492d3df2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.130795 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.138985 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg" (OuterVolumeSpecName: "kube-api-access-9l4pg") pod "c70d7dea-801a-4fe3-9b1f-5e37492d3df2" (UID: "c70d7dea-801a-4fe3-9b1f-5e37492d3df2"). InnerVolumeSpecName "kube-api-access-9l4pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.233868 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l4pg\" (UniqueName: \"kubernetes.io/projected/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-kube-api-access-9l4pg\") on node \"crc\" DevicePath \"\"" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.564339 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c70d7dea-801a-4fe3-9b1f-5e37492d3df2" (UID: "c70d7dea-801a-4fe3-9b1f-5e37492d3df2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.643501 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c70d7dea-801a-4fe3-9b1f-5e37492d3df2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.647917 4972 generic.go:334] "Generic (PLEG): container finished" podID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerID="9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a" exitCode=0 Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.647986 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerDied","Data":"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a"} Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.648023 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ctgz6" event={"ID":"c70d7dea-801a-4fe3-9b1f-5e37492d3df2","Type":"ContainerDied","Data":"95c431c167c8f6db7f8cd75b5b5a2bd742d8637ec3c1d466ae44da561e31e99e"} Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.648063 4972 scope.go:117] "RemoveContainer" containerID="9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.648318 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ctgz6" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.695760 4972 scope.go:117] "RemoveContainer" containerID="1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.699776 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.710323 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ctgz6"] Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.722829 4972 scope.go:117] "RemoveContainer" containerID="c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.774723 4972 scope.go:117] "RemoveContainer" containerID="9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a" Feb 28 11:12:19 crc kubenswrapper[4972]: E0228 11:12:19.775240 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a\": container with ID starting with 9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a not found: ID does not exist" containerID="9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.775285 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a"} err="failed to get container status \"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a\": rpc error: code = NotFound desc = could not find container \"9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a\": container with ID starting with 9904d4bf6709381f8ff203862fd730d8d0b336c06d54a7960bd9cec60d3c978a not found: ID does not exist" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.775317 4972 scope.go:117] "RemoveContainer" containerID="1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729" Feb 28 11:12:19 crc kubenswrapper[4972]: E0228 11:12:19.776194 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729\": container with ID starting with 1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729 not found: ID does not exist" containerID="1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.776229 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729"} err="failed to get container status \"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729\": rpc error: code = NotFound desc = could not find container \"1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729\": container with ID starting with 1c79decad128cdf6d1937beaa575f4c3a45c2e0b8d194d34e4a43cac4dffe729 not found: ID does not exist" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.776252 4972 scope.go:117] "RemoveContainer" containerID="c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64" Feb 28 11:12:19 crc kubenswrapper[4972]: E0228 11:12:19.776674 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64\": container with ID starting with c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64 not found: ID does not exist" containerID="c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.776699 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64"} err="failed to get container status \"c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64\": rpc error: code = NotFound desc = could not find container \"c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64\": container with ID starting with c11f13c971ea125cbe3dd28f3093106dec2a1823e245a1c15d1e9c17e6996c64 not found: ID does not exist" Feb 28 11:12:19 crc kubenswrapper[4972]: I0228 11:12:19.802508 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" path="/var/lib/kubelet/pods/c70d7dea-801a-4fe3-9b1f-5e37492d3df2/volumes" Feb 28 11:12:58 crc kubenswrapper[4972]: I0228 11:12:58.890952 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:12:58 crc kubenswrapper[4972]: I0228 11:12:58.891497 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:13:03 crc kubenswrapper[4972]: I0228 11:13:03.651961 4972 scope.go:117] "RemoveContainer" containerID="e756f91a1e08e3413bb3da9f0e0cffad199c75d096c6a43e9bd66c912ea894e0" Feb 28 11:13:28 crc kubenswrapper[4972]: I0228 11:13:28.891262 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:13:28 crc kubenswrapper[4972]: I0228 11:13:28.892619 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:13:58 crc kubenswrapper[4972]: I0228 11:13:58.890731 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:13:58 crc kubenswrapper[4972]: I0228 11:13:58.891824 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:13:58 crc kubenswrapper[4972]: I0228 11:13:58.891915 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:13:58 crc kubenswrapper[4972]: I0228 11:13:58.893193 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:13:58 crc kubenswrapper[4972]: I0228 11:13:58.893279 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c" gracePeriod=600 Feb 28 11:13:59 crc kubenswrapper[4972]: I0228 11:13:59.696700 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c" exitCode=0 Feb 28 11:13:59 crc kubenswrapper[4972]: I0228 11:13:59.696809 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c"} Feb 28 11:13:59 crc kubenswrapper[4972]: I0228 11:13:59.697848 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0"} Feb 28 11:13:59 crc kubenswrapper[4972]: I0228 11:13:59.697893 4972 scope.go:117] "RemoveContainer" containerID="0faeb5c6def7b098cd6c86b2e772259b877b6e5dddf8dc9580d37868dcae54c7" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.164947 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537954-6xn9d"] Feb 28 11:14:00 crc kubenswrapper[4972]: E0228 11:14:00.165690 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="extract-content" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.165715 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="extract-content" Feb 28 11:14:00 crc kubenswrapper[4972]: E0228 11:14:00.165751 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="extract-utilities" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.165765 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="extract-utilities" Feb 28 11:14:00 crc kubenswrapper[4972]: E0228 11:14:00.165792 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="registry-server" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.165804 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="registry-server" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.166108 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70d7dea-801a-4fe3-9b1f-5e37492d3df2" containerName="registry-server" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.167202 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.170150 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.170970 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.172510 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.179623 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537954-6xn9d"] Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.233615 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdbqh\" (UniqueName: \"kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh\") pod \"auto-csr-approver-29537954-6xn9d\" (UID: \"04322d21-316d-4235-9866-58ecdabedc3c\") " pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.336698 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdbqh\" (UniqueName: \"kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh\") pod \"auto-csr-approver-29537954-6xn9d\" (UID: \"04322d21-316d-4235-9866-58ecdabedc3c\") " pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.368301 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdbqh\" (UniqueName: \"kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh\") pod \"auto-csr-approver-29537954-6xn9d\" (UID: \"04322d21-316d-4235-9866-58ecdabedc3c\") " pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.497072 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:00 crc kubenswrapper[4972]: I0228 11:14:00.987808 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537954-6xn9d"] Feb 28 11:14:01 crc kubenswrapper[4972]: I0228 11:14:01.731646 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" event={"ID":"04322d21-316d-4235-9866-58ecdabedc3c","Type":"ContainerStarted","Data":"9b66870a486f893393c28b3477195155d6eb09f8ed22c447c6c6987cb5d449af"} Feb 28 11:14:02 crc kubenswrapper[4972]: I0228 11:14:02.744854 4972 generic.go:334] "Generic (PLEG): container finished" podID="04322d21-316d-4235-9866-58ecdabedc3c" containerID="647502943ba3db1a14f900a8e6b5275cd466122a8d2b9a7d973f17a032e3fa91" exitCode=0 Feb 28 11:14:02 crc kubenswrapper[4972]: I0228 11:14:02.744922 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" event={"ID":"04322d21-316d-4235-9866-58ecdabedc3c","Type":"ContainerDied","Data":"647502943ba3db1a14f900a8e6b5275cd466122a8d2b9a7d973f17a032e3fa91"} Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.122217 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.234382 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdbqh\" (UniqueName: \"kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh\") pod \"04322d21-316d-4235-9866-58ecdabedc3c\" (UID: \"04322d21-316d-4235-9866-58ecdabedc3c\") " Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.242493 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh" (OuterVolumeSpecName: "kube-api-access-pdbqh") pod "04322d21-316d-4235-9866-58ecdabedc3c" (UID: "04322d21-316d-4235-9866-58ecdabedc3c"). InnerVolumeSpecName "kube-api-access-pdbqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.339618 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdbqh\" (UniqueName: \"kubernetes.io/projected/04322d21-316d-4235-9866-58ecdabedc3c-kube-api-access-pdbqh\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.785717 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" event={"ID":"04322d21-316d-4235-9866-58ecdabedc3c","Type":"ContainerDied","Data":"9b66870a486f893393c28b3477195155d6eb09f8ed22c447c6c6987cb5d449af"} Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.785810 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b66870a486f893393c28b3477195155d6eb09f8ed22c447c6c6987cb5d449af" Feb 28 11:14:04 crc kubenswrapper[4972]: I0228 11:14:04.785820 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537954-6xn9d" Feb 28 11:14:05 crc kubenswrapper[4972]: I0228 11:14:05.209219 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537948-tpw5b"] Feb 28 11:14:05 crc kubenswrapper[4972]: I0228 11:14:05.217425 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537948-tpw5b"] Feb 28 11:14:05 crc kubenswrapper[4972]: I0228 11:14:05.807547 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ab2894-e898-444d-9586-40e9f2333332" path="/var/lib/kubelet/pods/00ab2894-e898-444d-9586-40e9f2333332/volumes" Feb 28 11:14:21 crc kubenswrapper[4972]: I0228 11:14:21.976558 4972 generic.go:334] "Generic (PLEG): container finished" podID="1a18fc65-40ad-474f-b28d-76e3d12cb6df" containerID="14fb18454c30a80416435bb8285f19d8407334dc50426c792ab856ee39fad4ac" exitCode=0 Feb 28 11:14:21 crc kubenswrapper[4972]: I0228 11:14:21.976658 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" event={"ID":"1a18fc65-40ad-474f-b28d-76e3d12cb6df","Type":"ContainerDied","Data":"14fb18454c30a80416435bb8285f19d8407334dc50426c792ab856ee39fad4ac"} Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.468954 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.638181 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0\") pod \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.638255 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory\") pod \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.638580 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam\") pod \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.638622 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle\") pod \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.638656 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824zw\" (UniqueName: \"kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw\") pod \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\" (UID: \"1a18fc65-40ad-474f-b28d-76e3d12cb6df\") " Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.647542 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw" (OuterVolumeSpecName: "kube-api-access-824zw") pod "1a18fc65-40ad-474f-b28d-76e3d12cb6df" (UID: "1a18fc65-40ad-474f-b28d-76e3d12cb6df"). InnerVolumeSpecName "kube-api-access-824zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.648278 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1a18fc65-40ad-474f-b28d-76e3d12cb6df" (UID: "1a18fc65-40ad-474f-b28d-76e3d12cb6df"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.683303 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "1a18fc65-40ad-474f-b28d-76e3d12cb6df" (UID: "1a18fc65-40ad-474f-b28d-76e3d12cb6df"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.688832 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory" (OuterVolumeSpecName: "inventory") pod "1a18fc65-40ad-474f-b28d-76e3d12cb6df" (UID: "1a18fc65-40ad-474f-b28d-76e3d12cb6df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.693114 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1a18fc65-40ad-474f-b28d-76e3d12cb6df" (UID: "1a18fc65-40ad-474f-b28d-76e3d12cb6df"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.741226 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.741263 4972 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.741273 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824zw\" (UniqueName: \"kubernetes.io/projected/1a18fc65-40ad-474f-b28d-76e3d12cb6df-kube-api-access-824zw\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.741283 4972 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:23 crc kubenswrapper[4972]: I0228 11:14:23.741292 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a18fc65-40ad-474f-b28d-76e3d12cb6df-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.004762 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" event={"ID":"1a18fc65-40ad-474f-b28d-76e3d12cb6df","Type":"ContainerDied","Data":"0edf7eec669c9e29c918c5d279e0e2c1a0519849ab084a04af4ee7e8dacd511a"} Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.005147 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0edf7eec669c9e29c918c5d279e0e2c1a0519849ab084a04af4ee7e8dacd511a" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.004909 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.194335 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd"] Feb 28 11:14:24 crc kubenswrapper[4972]: E0228 11:14:24.194739 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04322d21-316d-4235-9866-58ecdabedc3c" containerName="oc" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.194756 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="04322d21-316d-4235-9866-58ecdabedc3c" containerName="oc" Feb 28 11:14:24 crc kubenswrapper[4972]: E0228 11:14:24.194787 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a18fc65-40ad-474f-b28d-76e3d12cb6df" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.194795 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a18fc65-40ad-474f-b28d-76e3d12cb6df" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.195103 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a18fc65-40ad-474f-b28d-76e3d12cb6df" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.195147 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="04322d21-316d-4235-9866-58ecdabedc3c" containerName="oc" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.195842 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.201748 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.202013 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.202259 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.202572 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.202749 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.202915 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.203090 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.212350 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd"] Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.356894 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357074 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357165 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357238 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357326 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357524 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357586 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357809 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357856 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mc54\" (UniqueName: \"kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357914 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.357952 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460061 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460129 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460207 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460235 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mc54\" (UniqueName: \"kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460276 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460302 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460336 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460380 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460408 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460442 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.460516 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.461895 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.466609 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.467222 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.467500 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.467924 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.468286 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.468326 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.469118 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.470502 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.470832 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.479909 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mc54\" (UniqueName: \"kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dqvpd\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:24 crc kubenswrapper[4972]: I0228 11:14:24.523137 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:14:25 crc kubenswrapper[4972]: I0228 11:14:25.097701 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd"] Feb 28 11:14:26 crc kubenswrapper[4972]: I0228 11:14:26.025277 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" event={"ID":"d0655084-fd3f-4088-bb4e-755a5db445fe","Type":"ContainerStarted","Data":"249a6cfb72870868b89ddc729ed504713076c39aec47b8f2de36d7458bd139a0"} Feb 28 11:14:26 crc kubenswrapper[4972]: I0228 11:14:26.028211 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" event={"ID":"d0655084-fd3f-4088-bb4e-755a5db445fe","Type":"ContainerStarted","Data":"3a427dceb9f5648b341efab780bc3551f3ea4a97c78e2d73a9d47d8b73bf136b"} Feb 28 11:14:26 crc kubenswrapper[4972]: I0228 11:14:26.069927 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" podStartSLOduration=1.668179324 podStartE2EDuration="2.069890856s" podCreationTimestamp="2026-02-28 11:14:24 +0000 UTC" firstStartedPulling="2026-02-28 11:14:25.110727923 +0000 UTC m=+2322.022735661" lastFinishedPulling="2026-02-28 11:14:25.512439455 +0000 UTC m=+2322.424447193" observedRunningTime="2026-02-28 11:14:26.053580812 +0000 UTC m=+2322.965588580" watchObservedRunningTime="2026-02-28 11:14:26.069890856 +0000 UTC m=+2322.981898634" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.466390 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.468366 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.478938 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.552034 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.552106 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4skb8\" (UniqueName: \"kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.552146 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.653603 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.653663 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4skb8\" (UniqueName: \"kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.653699 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.654195 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.654243 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.672339 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4skb8\" (UniqueName: \"kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8\") pod \"community-operators-fznvt\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:27 crc kubenswrapper[4972]: I0228 11:14:27.789591 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:28 crc kubenswrapper[4972]: I0228 11:14:28.373855 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:29 crc kubenswrapper[4972]: I0228 11:14:29.057176 4972 generic.go:334] "Generic (PLEG): container finished" podID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerID="e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c" exitCode=0 Feb 28 11:14:29 crc kubenswrapper[4972]: I0228 11:14:29.057585 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerDied","Data":"e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c"} Feb 28 11:14:29 crc kubenswrapper[4972]: I0228 11:14:29.057611 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerStarted","Data":"f906717109842c3e828b4dc45b4de94b0cd67ccc45b82d239fb85634f1bd8568"} Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.073369 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerStarted","Data":"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe"} Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.491575 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.498245 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.516200 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.653335 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.653822 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szs4j\" (UniqueName: \"kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.654674 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.757349 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.757415 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.757496 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szs4j\" (UniqueName: \"kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.758266 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.758292 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.777905 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szs4j\" (UniqueName: \"kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j\") pod \"redhat-marketplace-9dgz4\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:30 crc kubenswrapper[4972]: I0228 11:14:30.826021 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:31 crc kubenswrapper[4972]: I0228 11:14:31.095971 4972 generic.go:334] "Generic (PLEG): container finished" podID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerID="11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe" exitCode=0 Feb 28 11:14:31 crc kubenswrapper[4972]: I0228 11:14:31.096058 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerDied","Data":"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe"} Feb 28 11:14:31 crc kubenswrapper[4972]: I0228 11:14:31.360228 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:32 crc kubenswrapper[4972]: I0228 11:14:32.117448 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerStarted","Data":"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e"} Feb 28 11:14:32 crc kubenswrapper[4972]: I0228 11:14:32.120904 4972 generic.go:334] "Generic (PLEG): container finished" podID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerID="4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338" exitCode=0 Feb 28 11:14:32 crc kubenswrapper[4972]: I0228 11:14:32.120999 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerDied","Data":"4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338"} Feb 28 11:14:32 crc kubenswrapper[4972]: I0228 11:14:32.121061 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerStarted","Data":"17fac43f0c09a434eb8a80170242751c45ab23faa7389b4369d05fb59a9e68a1"} Feb 28 11:14:32 crc kubenswrapper[4972]: I0228 11:14:32.163554 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fznvt" podStartSLOduration=2.728582915 podStartE2EDuration="5.16352223s" podCreationTimestamp="2026-02-28 11:14:27 +0000 UTC" firstStartedPulling="2026-02-28 11:14:29.059242609 +0000 UTC m=+2325.971250347" lastFinishedPulling="2026-02-28 11:14:31.494181914 +0000 UTC m=+2328.406189662" observedRunningTime="2026-02-28 11:14:32.143154976 +0000 UTC m=+2329.055162754" watchObservedRunningTime="2026-02-28 11:14:32.16352223 +0000 UTC m=+2329.075530008" Feb 28 11:14:33 crc kubenswrapper[4972]: I0228 11:14:33.131202 4972 generic.go:334] "Generic (PLEG): container finished" podID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerID="d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d" exitCode=0 Feb 28 11:14:33 crc kubenswrapper[4972]: I0228 11:14:33.132721 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerDied","Data":"d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d"} Feb 28 11:14:33 crc kubenswrapper[4972]: E0228 11:14:33.199883 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93cf9b91_7433_459b_a36e_6d150c260aa3.slice/crio-conmon-d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d.scope\": RecentStats: unable to find data in memory cache]" Feb 28 11:14:34 crc kubenswrapper[4972]: I0228 11:14:34.152619 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerStarted","Data":"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f"} Feb 28 11:14:34 crc kubenswrapper[4972]: I0228 11:14:34.195949 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9dgz4" podStartSLOduration=2.822769822 podStartE2EDuration="4.195907921s" podCreationTimestamp="2026-02-28 11:14:30 +0000 UTC" firstStartedPulling="2026-02-28 11:14:32.127240863 +0000 UTC m=+2329.039248651" lastFinishedPulling="2026-02-28 11:14:33.500379022 +0000 UTC m=+2330.412386750" observedRunningTime="2026-02-28 11:14:34.182364122 +0000 UTC m=+2331.094371900" watchObservedRunningTime="2026-02-28 11:14:34.195907921 +0000 UTC m=+2331.107915679" Feb 28 11:14:37 crc kubenswrapper[4972]: I0228 11:14:37.806924 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:37 crc kubenswrapper[4972]: I0228 11:14:37.806982 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:37 crc kubenswrapper[4972]: I0228 11:14:37.861502 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:38 crc kubenswrapper[4972]: I0228 11:14:38.286850 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:40 crc kubenswrapper[4972]: I0228 11:14:40.826630 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:40 crc kubenswrapper[4972]: I0228 11:14:40.827060 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:40 crc kubenswrapper[4972]: I0228 11:14:40.918070 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:41 crc kubenswrapper[4972]: I0228 11:14:41.340076 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:42 crc kubenswrapper[4972]: I0228 11:14:42.267488 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:42 crc kubenswrapper[4972]: I0228 11:14:42.656743 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:42 crc kubenswrapper[4972]: I0228 11:14:42.657016 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fznvt" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="registry-server" containerID="cri-o://eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e" gracePeriod=2 Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.141925 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.184401 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities\") pod \"a143f056-89ba-4a30-a7ca-cc36e459d178\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.184570 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content\") pod \"a143f056-89ba-4a30-a7ca-cc36e459d178\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.184633 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4skb8\" (UniqueName: \"kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8\") pod \"a143f056-89ba-4a30-a7ca-cc36e459d178\" (UID: \"a143f056-89ba-4a30-a7ca-cc36e459d178\") " Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.186011 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities" (OuterVolumeSpecName: "utilities") pod "a143f056-89ba-4a30-a7ca-cc36e459d178" (UID: "a143f056-89ba-4a30-a7ca-cc36e459d178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.193216 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8" (OuterVolumeSpecName: "kube-api-access-4skb8") pod "a143f056-89ba-4a30-a7ca-cc36e459d178" (UID: "a143f056-89ba-4a30-a7ca-cc36e459d178"). InnerVolumeSpecName "kube-api-access-4skb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.250643 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a143f056-89ba-4a30-a7ca-cc36e459d178" (UID: "a143f056-89ba-4a30-a7ca-cc36e459d178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.278967 4972 generic.go:334] "Generic (PLEG): container finished" podID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerID="eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e" exitCode=0 Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.279064 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerDied","Data":"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e"} Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.279085 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fznvt" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.279114 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fznvt" event={"ID":"a143f056-89ba-4a30-a7ca-cc36e459d178","Type":"ContainerDied","Data":"f906717109842c3e828b4dc45b4de94b0cd67ccc45b82d239fb85634f1bd8568"} Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.279140 4972 scope.go:117] "RemoveContainer" containerID="eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.286899 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.286941 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a143f056-89ba-4a30-a7ca-cc36e459d178-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.286954 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4skb8\" (UniqueName: \"kubernetes.io/projected/a143f056-89ba-4a30-a7ca-cc36e459d178-kube-api-access-4skb8\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.322989 4972 scope.go:117] "RemoveContainer" containerID="11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.331406 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.345183 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fznvt"] Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.351726 4972 scope.go:117] "RemoveContainer" containerID="e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.382786 4972 scope.go:117] "RemoveContainer" containerID="eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e" Feb 28 11:14:43 crc kubenswrapper[4972]: E0228 11:14:43.383908 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e\": container with ID starting with eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e not found: ID does not exist" containerID="eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.383971 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e"} err="failed to get container status \"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e\": rpc error: code = NotFound desc = could not find container \"eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e\": container with ID starting with eef98fa062af9a68fab18f0dba633e68fc90ce83143917203dabf04c802aef3e not found: ID does not exist" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.384010 4972 scope.go:117] "RemoveContainer" containerID="11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe" Feb 28 11:14:43 crc kubenswrapper[4972]: E0228 11:14:43.384394 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe\": container with ID starting with 11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe not found: ID does not exist" containerID="11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.384534 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe"} err="failed to get container status \"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe\": rpc error: code = NotFound desc = could not find container \"11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe\": container with ID starting with 11cc5a9fcdd98759eae856132ce880143a6f53117e6302cb3ebbc8d768bb8afe not found: ID does not exist" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.384655 4972 scope.go:117] "RemoveContainer" containerID="e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c" Feb 28 11:14:43 crc kubenswrapper[4972]: E0228 11:14:43.385070 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c\": container with ID starting with e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c not found: ID does not exist" containerID="e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.385182 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c"} err="failed to get container status \"e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c\": rpc error: code = NotFound desc = could not find container \"e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c\": container with ID starting with e27f4cdba66697029c5f087e93f4355d17efff7084269cd615b3b5b4fe74695c not found: ID does not exist" Feb 28 11:14:43 crc kubenswrapper[4972]: E0228 11:14:43.460729 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda143f056_89ba_4a30_a7ca_cc36e459d178.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda143f056_89ba_4a30_a7ca_cc36e459d178.slice/crio-f906717109842c3e828b4dc45b4de94b0cd67ccc45b82d239fb85634f1bd8568\": RecentStats: unable to find data in memory cache]" Feb 28 11:14:43 crc kubenswrapper[4972]: I0228 11:14:43.804034 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" path="/var/lib/kubelet/pods/a143f056-89ba-4a30-a7ca-cc36e459d178/volumes" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.291148 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9dgz4" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="registry-server" containerID="cri-o://4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f" gracePeriod=2 Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.806634 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.825054 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities\") pod \"93cf9b91-7433-459b-a36e-6d150c260aa3\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.825177 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content\") pod \"93cf9b91-7433-459b-a36e-6d150c260aa3\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.825344 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szs4j\" (UniqueName: \"kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j\") pod \"93cf9b91-7433-459b-a36e-6d150c260aa3\" (UID: \"93cf9b91-7433-459b-a36e-6d150c260aa3\") " Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.825761 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities" (OuterVolumeSpecName: "utilities") pod "93cf9b91-7433-459b-a36e-6d150c260aa3" (UID: "93cf9b91-7433-459b-a36e-6d150c260aa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.826299 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.839000 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j" (OuterVolumeSpecName: "kube-api-access-szs4j") pod "93cf9b91-7433-459b-a36e-6d150c260aa3" (UID: "93cf9b91-7433-459b-a36e-6d150c260aa3"). InnerVolumeSpecName "kube-api-access-szs4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.850652 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93cf9b91-7433-459b-a36e-6d150c260aa3" (UID: "93cf9b91-7433-459b-a36e-6d150c260aa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.928296 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szs4j\" (UniqueName: \"kubernetes.io/projected/93cf9b91-7433-459b-a36e-6d150c260aa3-kube-api-access-szs4j\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:44 crc kubenswrapper[4972]: I0228 11:14:44.928564 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93cf9b91-7433-459b-a36e-6d150c260aa3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.306411 4972 generic.go:334] "Generic (PLEG): container finished" podID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerID="4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f" exitCode=0 Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.306501 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dgz4" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.306519 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerDied","Data":"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f"} Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.306575 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dgz4" event={"ID":"93cf9b91-7433-459b-a36e-6d150c260aa3","Type":"ContainerDied","Data":"17fac43f0c09a434eb8a80170242751c45ab23faa7389b4369d05fb59a9e68a1"} Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.306602 4972 scope.go:117] "RemoveContainer" containerID="4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.334876 4972 scope.go:117] "RemoveContainer" containerID="d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.353900 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.377012 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dgz4"] Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.394257 4972 scope.go:117] "RemoveContainer" containerID="4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.427502 4972 scope.go:117] "RemoveContainer" containerID="4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f" Feb 28 11:14:45 crc kubenswrapper[4972]: E0228 11:14:45.427756 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f\": container with ID starting with 4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f not found: ID does not exist" containerID="4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.427810 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f"} err="failed to get container status \"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f\": rpc error: code = NotFound desc = could not find container \"4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f\": container with ID starting with 4d55fe1e91c888e18180511f85d5c42cd8294838abf81bf45eb9691c92e1b59f not found: ID does not exist" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.427842 4972 scope.go:117] "RemoveContainer" containerID="d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d" Feb 28 11:14:45 crc kubenswrapper[4972]: E0228 11:14:45.428241 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d\": container with ID starting with d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d not found: ID does not exist" containerID="d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.428282 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d"} err="failed to get container status \"d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d\": rpc error: code = NotFound desc = could not find container \"d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d\": container with ID starting with d65a93293cbed6d5ae7fcfb5e8bdf33cba0c841c2fccbfbe864c5d5d709a646d not found: ID does not exist" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.428308 4972 scope.go:117] "RemoveContainer" containerID="4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338" Feb 28 11:14:45 crc kubenswrapper[4972]: E0228 11:14:45.428565 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338\": container with ID starting with 4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338 not found: ID does not exist" containerID="4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.428596 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338"} err="failed to get container status \"4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338\": rpc error: code = NotFound desc = could not find container \"4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338\": container with ID starting with 4fbfcfc0e80378554ddafe515cb00ca04bfb708d4059147c238e8f2bab768338 not found: ID does not exist" Feb 28 11:14:45 crc kubenswrapper[4972]: I0228 11:14:45.802273 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" path="/var/lib/kubelet/pods/93cf9b91-7433-459b-a36e-6d150c260aa3/volumes" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.173999 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq"] Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175156 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175179 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175212 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="extract-utilities" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175222 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="extract-utilities" Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175241 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175248 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175266 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="extract-utilities" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175274 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="extract-utilities" Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175293 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="extract-content" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175302 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="extract-content" Feb 28 11:15:00 crc kubenswrapper[4972]: E0228 11:15:00.175345 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="extract-content" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175355 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="extract-content" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175604 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="93cf9b91-7433-459b-a36e-6d150c260aa3" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.175628 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a143f056-89ba-4a30-a7ca-cc36e459d178" containerName="registry-server" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.176437 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.179887 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.181806 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.187962 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq"] Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.307543 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.307771 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpz7f\" (UniqueName: \"kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.307819 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.410575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpz7f\" (UniqueName: \"kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.410677 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.410816 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.411802 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.421723 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.439728 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpz7f\" (UniqueName: \"kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f\") pod \"collect-profiles-29537955-g7hkq\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:00 crc kubenswrapper[4972]: I0228 11:15:00.518212 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:01 crc kubenswrapper[4972]: I0228 11:15:01.031353 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq"] Feb 28 11:15:01 crc kubenswrapper[4972]: I0228 11:15:01.545247 4972 generic.go:334] "Generic (PLEG): container finished" podID="622493aa-a52e-46a2-89a0-c8d387a4a6f9" containerID="0a90943728fc8cfbda954a728b39214a2b0a35d039c66493182208d728b3233e" exitCode=0 Feb 28 11:15:01 crc kubenswrapper[4972]: I0228 11:15:01.545309 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" event={"ID":"622493aa-a52e-46a2-89a0-c8d387a4a6f9","Type":"ContainerDied","Data":"0a90943728fc8cfbda954a728b39214a2b0a35d039c66493182208d728b3233e"} Feb 28 11:15:01 crc kubenswrapper[4972]: I0228 11:15:01.546469 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" event={"ID":"622493aa-a52e-46a2-89a0-c8d387a4a6f9","Type":"ContainerStarted","Data":"319c9adda798925f0563c6b970b93ea5af8435283cb9eccf93c46f5178710b92"} Feb 28 11:15:02 crc kubenswrapper[4972]: I0228 11:15:02.959500 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.083845 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpz7f\" (UniqueName: \"kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f\") pod \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.084012 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume\") pod \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.084211 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume\") pod \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\" (UID: \"622493aa-a52e-46a2-89a0-c8d387a4a6f9\") " Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.085130 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume" (OuterVolumeSpecName: "config-volume") pod "622493aa-a52e-46a2-89a0-c8d387a4a6f9" (UID: "622493aa-a52e-46a2-89a0-c8d387a4a6f9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.092631 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "622493aa-a52e-46a2-89a0-c8d387a4a6f9" (UID: "622493aa-a52e-46a2-89a0-c8d387a4a6f9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.093328 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f" (OuterVolumeSpecName: "kube-api-access-fpz7f") pod "622493aa-a52e-46a2-89a0-c8d387a4a6f9" (UID: "622493aa-a52e-46a2-89a0-c8d387a4a6f9"). InnerVolumeSpecName "kube-api-access-fpz7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.186713 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/622493aa-a52e-46a2-89a0-c8d387a4a6f9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.186748 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpz7f\" (UniqueName: \"kubernetes.io/projected/622493aa-a52e-46a2-89a0-c8d387a4a6f9-kube-api-access-fpz7f\") on node \"crc\" DevicePath \"\"" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.186758 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/622493aa-a52e-46a2-89a0-c8d387a4a6f9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.574964 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" event={"ID":"622493aa-a52e-46a2-89a0-c8d387a4a6f9","Type":"ContainerDied","Data":"319c9adda798925f0563c6b970b93ea5af8435283cb9eccf93c46f5178710b92"} Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.575021 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319c9adda798925f0563c6b970b93ea5af8435283cb9eccf93c46f5178710b92" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.575088 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537955-g7hkq" Feb 28 11:15:03 crc kubenswrapper[4972]: I0228 11:15:03.819364 4972 scope.go:117] "RemoveContainer" containerID="a0adbf6be0ff881a57e0c3d071546747690b88ce1f3c51f98220d088cbbf18a4" Feb 28 11:15:04 crc kubenswrapper[4972]: I0228 11:15:04.048182 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48"] Feb 28 11:15:04 crc kubenswrapper[4972]: I0228 11:15:04.056407 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537910-gdj48"] Feb 28 11:15:05 crc kubenswrapper[4972]: I0228 11:15:05.813228 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde838bc-7c41-4da3-a00d-bb704e436faf" path="/var/lib/kubelet/pods/dde838bc-7c41-4da3-a00d-bb704e436faf/volumes" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.182625 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537956-zwpgk"] Feb 28 11:16:00 crc kubenswrapper[4972]: E0228 11:16:00.187361 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622493aa-a52e-46a2-89a0-c8d387a4a6f9" containerName="collect-profiles" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.187579 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="622493aa-a52e-46a2-89a0-c8d387a4a6f9" containerName="collect-profiles" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.188326 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="622493aa-a52e-46a2-89a0-c8d387a4a6f9" containerName="collect-profiles" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.189860 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.196582 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.197977 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.198173 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.229588 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537956-zwpgk"] Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.304707 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nhdc\" (UniqueName: \"kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc\") pod \"auto-csr-approver-29537956-zwpgk\" (UID: \"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f\") " pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.407085 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nhdc\" (UniqueName: \"kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc\") pod \"auto-csr-approver-29537956-zwpgk\" (UID: \"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f\") " pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.446390 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nhdc\" (UniqueName: \"kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc\") pod \"auto-csr-approver-29537956-zwpgk\" (UID: \"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f\") " pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:00 crc kubenswrapper[4972]: I0228 11:16:00.531147 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:01 crc kubenswrapper[4972]: I0228 11:16:01.111593 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537956-zwpgk"] Feb 28 11:16:01 crc kubenswrapper[4972]: I0228 11:16:01.125665 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:16:01 crc kubenswrapper[4972]: I0228 11:16:01.276291 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" event={"ID":"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f","Type":"ContainerStarted","Data":"ec92ee7fe619fbaddf67e4b57105e6188affeefbe66146bd5ab5187c49cbd8a7"} Feb 28 11:16:03 crc kubenswrapper[4972]: I0228 11:16:03.307189 4972 generic.go:334] "Generic (PLEG): container finished" podID="9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" containerID="bf80b48e1056cf9f4b6476b37554a153eb6a522ce301d7ac1395479ffc334599" exitCode=0 Feb 28 11:16:03 crc kubenswrapper[4972]: I0228 11:16:03.307318 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" event={"ID":"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f","Type":"ContainerDied","Data":"bf80b48e1056cf9f4b6476b37554a153eb6a522ce301d7ac1395479ffc334599"} Feb 28 11:16:03 crc kubenswrapper[4972]: I0228 11:16:03.953411 4972 scope.go:117] "RemoveContainer" containerID="6788e52a2d1f4fde8c8e40b952f5901a77d5f4effd4e76d1e7d77aace5dbee44" Feb 28 11:16:04 crc kubenswrapper[4972]: I0228 11:16:04.731234 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:04 crc kubenswrapper[4972]: I0228 11:16:04.737757 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nhdc\" (UniqueName: \"kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc\") pod \"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f\" (UID: \"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f\") " Feb 28 11:16:04 crc kubenswrapper[4972]: I0228 11:16:04.747639 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc" (OuterVolumeSpecName: "kube-api-access-6nhdc") pod "9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" (UID: "9f18cd6e-2519-4ee0-bcf7-9ade01703f7f"). InnerVolumeSpecName "kube-api-access-6nhdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:16:04 crc kubenswrapper[4972]: I0228 11:16:04.840583 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nhdc\" (UniqueName: \"kubernetes.io/projected/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f-kube-api-access-6nhdc\") on node \"crc\" DevicePath \"\"" Feb 28 11:16:05 crc kubenswrapper[4972]: I0228 11:16:05.344089 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" event={"ID":"9f18cd6e-2519-4ee0-bcf7-9ade01703f7f","Type":"ContainerDied","Data":"ec92ee7fe619fbaddf67e4b57105e6188affeefbe66146bd5ab5187c49cbd8a7"} Feb 28 11:16:05 crc kubenswrapper[4972]: I0228 11:16:05.344145 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec92ee7fe619fbaddf67e4b57105e6188affeefbe66146bd5ab5187c49cbd8a7" Feb 28 11:16:05 crc kubenswrapper[4972]: I0228 11:16:05.344158 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537956-zwpgk" Feb 28 11:16:05 crc kubenswrapper[4972]: E0228 11:16:05.559739 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f18cd6e_2519_4ee0_bcf7_9ade01703f7f.slice/crio-ec92ee7fe619fbaddf67e4b57105e6188affeefbe66146bd5ab5187c49cbd8a7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f18cd6e_2519_4ee0_bcf7_9ade01703f7f.slice\": RecentStats: unable to find data in memory cache]" Feb 28 11:16:05 crc kubenswrapper[4972]: I0228 11:16:05.818027 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537950-r9rpl"] Feb 28 11:16:05 crc kubenswrapper[4972]: I0228 11:16:05.830601 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537950-r9rpl"] Feb 28 11:16:07 crc kubenswrapper[4972]: I0228 11:16:07.849784 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37f24d57-f1a3-4785-b4be-54d73bbde481" path="/var/lib/kubelet/pods/37f24d57-f1a3-4785-b4be-54d73bbde481/volumes" Feb 28 11:16:28 crc kubenswrapper[4972]: I0228 11:16:28.890927 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:16:28 crc kubenswrapper[4972]: I0228 11:16:28.891811 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:16:58 crc kubenswrapper[4972]: I0228 11:16:58.890659 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:16:58 crc kubenswrapper[4972]: I0228 11:16:58.891829 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:17:00 crc kubenswrapper[4972]: I0228 11:17:00.102237 4972 generic.go:334] "Generic (PLEG): container finished" podID="d0655084-fd3f-4088-bb4e-755a5db445fe" containerID="249a6cfb72870868b89ddc729ed504713076c39aec47b8f2de36d7458bd139a0" exitCode=0 Feb 28 11:17:00 crc kubenswrapper[4972]: I0228 11:17:00.102331 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" event={"ID":"d0655084-fd3f-4088-bb4e-755a5db445fe","Type":"ContainerDied","Data":"249a6cfb72870868b89ddc729ed504713076c39aec47b8f2de36d7458bd139a0"} Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.638632 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676408 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676637 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676699 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676840 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676890 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.676948 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.677135 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mc54\" (UniqueName: \"kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.677229 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.677295 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.677370 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.677437 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1\") pod \"d0655084-fd3f-4088-bb4e-755a5db445fe\" (UID: \"d0655084-fd3f-4088-bb4e-755a5db445fe\") " Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.692411 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.701821 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54" (OuterVolumeSpecName: "kube-api-access-6mc54") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "kube-api-access-6mc54". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.751808 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory" (OuterVolumeSpecName: "inventory") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.752599 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.754869 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.763822 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.771122 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.785369 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.789714 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.791941 4972 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.791983 4972 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.791993 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mc54\" (UniqueName: \"kubernetes.io/projected/d0655084-fd3f-4088-bb4e-755a5db445fe-kube-api-access-6mc54\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792005 4972 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792016 4972 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792029 4972 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792038 4972 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792048 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.792058 4972 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.798044 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.818431 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d0655084-fd3f-4088-bb4e-755a5db445fe" (UID: "d0655084-fd3f-4088-bb4e-755a5db445fe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.893919 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:01 crc kubenswrapper[4972]: I0228 11:17:01.893968 4972 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d0655084-fd3f-4088-bb4e-755a5db445fe-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.131596 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" event={"ID":"d0655084-fd3f-4088-bb4e-755a5db445fe","Type":"ContainerDied","Data":"3a427dceb9f5648b341efab780bc3551f3ea4a97c78e2d73a9d47d8b73bf136b"} Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.131969 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a427dceb9f5648b341efab780bc3551f3ea4a97c78e2d73a9d47d8b73bf136b" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.132057 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dqvpd" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.366891 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f"] Feb 28 11:17:02 crc kubenswrapper[4972]: E0228 11:17:02.367354 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" containerName="oc" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.367371 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" containerName="oc" Feb 28 11:17:02 crc kubenswrapper[4972]: E0228 11:17:02.367393 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0655084-fd3f-4088-bb4e-755a5db445fe" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.367402 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0655084-fd3f-4088-bb4e-755a5db445fe" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.367609 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" containerName="oc" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.367629 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0655084-fd3f-4088-bb4e-755a5db445fe" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.368291 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.373116 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.373293 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.373302 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sb92d" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.373420 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.373541 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.394525 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f"] Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.505726 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.505787 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.505867 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.505889 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s86jx\" (UniqueName: \"kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.505909 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.506117 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.506194 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.608893 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.609296 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.609622 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.609794 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s86jx\" (UniqueName: \"kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.609957 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.610155 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.610313 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.615960 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.616643 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.616920 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.617426 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.629791 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.634404 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.640178 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s86jx\" (UniqueName: \"kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-shq6f\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:02 crc kubenswrapper[4972]: I0228 11:17:02.687610 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:17:03 crc kubenswrapper[4972]: I0228 11:17:03.345836 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f"] Feb 28 11:17:04 crc kubenswrapper[4972]: I0228 11:17:04.053417 4972 scope.go:117] "RemoveContainer" containerID="18e9497d2508fbbf53b5cef655b74f241514b6313ef72e2a3eb39a7962001565" Feb 28 11:17:04 crc kubenswrapper[4972]: I0228 11:17:04.171182 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" event={"ID":"a667491e-8728-4640-86b9-410f2cf770c9","Type":"ContainerStarted","Data":"642a411b147c45383b46447b4b2e66359cebbb2a02efe8bbba5abfd14ab7109a"} Feb 28 11:17:04 crc kubenswrapper[4972]: I0228 11:17:04.171244 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" event={"ID":"a667491e-8728-4640-86b9-410f2cf770c9","Type":"ContainerStarted","Data":"d253b20dd765b35654f0c26179a0f8dd4ef0f156026eaf073f13495923064829"} Feb 28 11:17:04 crc kubenswrapper[4972]: I0228 11:17:04.195433 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" podStartSLOduration=1.707198166 podStartE2EDuration="2.195404758s" podCreationTimestamp="2026-02-28 11:17:02 +0000 UTC" firstStartedPulling="2026-02-28 11:17:03.356342733 +0000 UTC m=+2480.268350481" lastFinishedPulling="2026-02-28 11:17:03.844549305 +0000 UTC m=+2480.756557073" observedRunningTime="2026-02-28 11:17:04.192742978 +0000 UTC m=+2481.104750756" watchObservedRunningTime="2026-02-28 11:17:04.195404758 +0000 UTC m=+2481.107412496" Feb 28 11:17:28 crc kubenswrapper[4972]: I0228 11:17:28.890900 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:17:28 crc kubenswrapper[4972]: I0228 11:17:28.891935 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:17:28 crc kubenswrapper[4972]: I0228 11:17:28.892016 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:17:28 crc kubenswrapper[4972]: I0228 11:17:28.893567 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:17:28 crc kubenswrapper[4972]: I0228 11:17:28.893678 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" gracePeriod=600 Feb 28 11:17:29 crc kubenswrapper[4972]: E0228 11:17:29.048989 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:17:29 crc kubenswrapper[4972]: I0228 11:17:29.519421 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" exitCode=0 Feb 28 11:17:29 crc kubenswrapper[4972]: I0228 11:17:29.519983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0"} Feb 28 11:17:29 crc kubenswrapper[4972]: I0228 11:17:29.520183 4972 scope.go:117] "RemoveContainer" containerID="fe3866852118bce726eda12d23d8d138f18b9eba6706e6781bad336de13fcc4c" Feb 28 11:17:29 crc kubenswrapper[4972]: I0228 11:17:29.521247 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:17:29 crc kubenswrapper[4972]: E0228 11:17:29.521754 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:17:42 crc kubenswrapper[4972]: I0228 11:17:42.789485 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:17:42 crc kubenswrapper[4972]: E0228 11:17:42.790768 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:17:53 crc kubenswrapper[4972]: I0228 11:17:53.811486 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:17:53 crc kubenswrapper[4972]: E0228 11:17:53.812540 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.178093 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537958-thnxn"] Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.182085 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.185328 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.185845 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.188754 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.192052 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537958-thnxn"] Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.249943 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gndfv\" (UniqueName: \"kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv\") pod \"auto-csr-approver-29537958-thnxn\" (UID: \"2ccb883b-1ed6-4dfc-a564-82a65b472b1f\") " pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.353213 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gndfv\" (UniqueName: \"kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv\") pod \"auto-csr-approver-29537958-thnxn\" (UID: \"2ccb883b-1ed6-4dfc-a564-82a65b472b1f\") " pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.393782 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gndfv\" (UniqueName: \"kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv\") pod \"auto-csr-approver-29537958-thnxn\" (UID: \"2ccb883b-1ed6-4dfc-a564-82a65b472b1f\") " pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:00 crc kubenswrapper[4972]: I0228 11:18:00.523590 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:01 crc kubenswrapper[4972]: I0228 11:18:01.068122 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537958-thnxn"] Feb 28 11:18:01 crc kubenswrapper[4972]: I0228 11:18:01.950773 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537958-thnxn" event={"ID":"2ccb883b-1ed6-4dfc-a564-82a65b472b1f","Type":"ContainerStarted","Data":"fb79c8ecc4729438013288c003fe04969b0c11504695109929de6e422f195d6a"} Feb 28 11:18:02 crc kubenswrapper[4972]: I0228 11:18:02.964906 4972 generic.go:334] "Generic (PLEG): container finished" podID="2ccb883b-1ed6-4dfc-a564-82a65b472b1f" containerID="d0649da349332a55de361d852164f2550873209b3a8e4a4f5b35b13f312b6d34" exitCode=0 Feb 28 11:18:02 crc kubenswrapper[4972]: I0228 11:18:02.964993 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537958-thnxn" event={"ID":"2ccb883b-1ed6-4dfc-a564-82a65b472b1f","Type":"ContainerDied","Data":"d0649da349332a55de361d852164f2550873209b3a8e4a4f5b35b13f312b6d34"} Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.409143 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.584189 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gndfv\" (UniqueName: \"kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv\") pod \"2ccb883b-1ed6-4dfc-a564-82a65b472b1f\" (UID: \"2ccb883b-1ed6-4dfc-a564-82a65b472b1f\") " Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.598974 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv" (OuterVolumeSpecName: "kube-api-access-gndfv") pod "2ccb883b-1ed6-4dfc-a564-82a65b472b1f" (UID: "2ccb883b-1ed6-4dfc-a564-82a65b472b1f"). InnerVolumeSpecName "kube-api-access-gndfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.688348 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gndfv\" (UniqueName: \"kubernetes.io/projected/2ccb883b-1ed6-4dfc-a564-82a65b472b1f-kube-api-access-gndfv\") on node \"crc\" DevicePath \"\"" Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.992690 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537958-thnxn" event={"ID":"2ccb883b-1ed6-4dfc-a564-82a65b472b1f","Type":"ContainerDied","Data":"fb79c8ecc4729438013288c003fe04969b0c11504695109929de6e422f195d6a"} Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.992765 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537958-thnxn" Feb 28 11:18:04 crc kubenswrapper[4972]: I0228 11:18:04.992775 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb79c8ecc4729438013288c003fe04969b0c11504695109929de6e422f195d6a" Feb 28 11:18:05 crc kubenswrapper[4972]: I0228 11:18:05.496318 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537952-rnxj8"] Feb 28 11:18:05 crc kubenswrapper[4972]: I0228 11:18:05.503884 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537952-rnxj8"] Feb 28 11:18:05 crc kubenswrapper[4972]: I0228 11:18:05.808971 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cfa36c3-a86e-4fad-b923-e73fbfb34e83" path="/var/lib/kubelet/pods/7cfa36c3-a86e-4fad-b923-e73fbfb34e83/volumes" Feb 28 11:18:07 crc kubenswrapper[4972]: I0228 11:18:07.792666 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:18:07 crc kubenswrapper[4972]: E0228 11:18:07.793137 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:18:18 crc kubenswrapper[4972]: I0228 11:18:18.790098 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:18:18 crc kubenswrapper[4972]: E0228 11:18:18.791612 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:18:32 crc kubenswrapper[4972]: I0228 11:18:32.789902 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:18:32 crc kubenswrapper[4972]: E0228 11:18:32.791110 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:18:46 crc kubenswrapper[4972]: I0228 11:18:46.790221 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:18:46 crc kubenswrapper[4972]: E0228 11:18:46.791586 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:18:57 crc kubenswrapper[4972]: I0228 11:18:57.791992 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:18:57 crc kubenswrapper[4972]: E0228 11:18:57.793405 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:19:04 crc kubenswrapper[4972]: I0228 11:19:04.222060 4972 scope.go:117] "RemoveContainer" containerID="2a2ba8b8e8245bb9874abfc50b011576b73ab63d1908d2d191e7d5969f9d0b98" Feb 28 11:19:08 crc kubenswrapper[4972]: I0228 11:19:08.789515 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:19:08 crc kubenswrapper[4972]: E0228 11:19:08.790763 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:19:21 crc kubenswrapper[4972]: I0228 11:19:21.790941 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:19:21 crc kubenswrapper[4972]: E0228 11:19:21.792925 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:19:33 crc kubenswrapper[4972]: I0228 11:19:33.802838 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:19:33 crc kubenswrapper[4972]: E0228 11:19:33.804537 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:19:44 crc kubenswrapper[4972]: I0228 11:19:44.285749 4972 generic.go:334] "Generic (PLEG): container finished" podID="a667491e-8728-4640-86b9-410f2cf770c9" containerID="642a411b147c45383b46447b4b2e66359cebbb2a02efe8bbba5abfd14ab7109a" exitCode=0 Feb 28 11:19:44 crc kubenswrapper[4972]: I0228 11:19:44.285911 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" event={"ID":"a667491e-8728-4640-86b9-410f2cf770c9","Type":"ContainerDied","Data":"642a411b147c45383b46447b4b2e66359cebbb2a02efe8bbba5abfd14ab7109a"} Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.790642 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:19:45 crc kubenswrapper[4972]: E0228 11:19:45.791524 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.850668 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.972427 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.972634 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.972750 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.972901 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.973019 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.973081 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.973221 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s86jx\" (UniqueName: \"kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx\") pod \"a667491e-8728-4640-86b9-410f2cf770c9\" (UID: \"a667491e-8728-4640-86b9-410f2cf770c9\") " Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.981418 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:45 crc kubenswrapper[4972]: I0228 11:19:45.983145 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx" (OuterVolumeSpecName: "kube-api-access-s86jx") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "kube-api-access-s86jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.005679 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.005928 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.009004 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.009063 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory" (OuterVolumeSpecName: "inventory") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.016211 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a667491e-8728-4640-86b9-410f2cf770c9" (UID: "a667491e-8728-4640-86b9-410f2cf770c9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.077097 4972 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.078312 4972 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.078492 4972 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.078600 4972 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.078685 4972 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-inventory\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.079267 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s86jx\" (UniqueName: \"kubernetes.io/projected/a667491e-8728-4640-86b9-410f2cf770c9-kube-api-access-s86jx\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.079496 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a667491e-8728-4640-86b9-410f2cf770c9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.317143 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" event={"ID":"a667491e-8728-4640-86b9-410f2cf770c9","Type":"ContainerDied","Data":"d253b20dd765b35654f0c26179a0f8dd4ef0f156026eaf073f13495923064829"} Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.317210 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-shq6f" Feb 28 11:19:46 crc kubenswrapper[4972]: I0228 11:19:46.317229 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d253b20dd765b35654f0c26179a0f8dd4ef0f156026eaf073f13495923064829" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.162610 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537960-pgr9r"] Feb 28 11:20:00 crc kubenswrapper[4972]: E0228 11:20:00.164169 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ccb883b-1ed6-4dfc-a564-82a65b472b1f" containerName="oc" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.164194 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ccb883b-1ed6-4dfc-a564-82a65b472b1f" containerName="oc" Feb 28 11:20:00 crc kubenswrapper[4972]: E0228 11:20:00.164248 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a667491e-8728-4640-86b9-410f2cf770c9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.164263 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a667491e-8728-4640-86b9-410f2cf770c9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.164682 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ccb883b-1ed6-4dfc-a564-82a65b472b1f" containerName="oc" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.164752 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a667491e-8728-4640-86b9-410f2cf770c9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.165885 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.170292 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.171339 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.176278 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.178851 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537960-pgr9r"] Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.202575 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mrr\" (UniqueName: \"kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr\") pod \"auto-csr-approver-29537960-pgr9r\" (UID: \"f79393b1-d156-4757-aaff-8b6fb64a57d4\") " pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.305442 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mrr\" (UniqueName: \"kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr\") pod \"auto-csr-approver-29537960-pgr9r\" (UID: \"f79393b1-d156-4757-aaff-8b6fb64a57d4\") " pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.351889 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mrr\" (UniqueName: \"kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr\") pod \"auto-csr-approver-29537960-pgr9r\" (UID: \"f79393b1-d156-4757-aaff-8b6fb64a57d4\") " pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.495035 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.789898 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:20:00 crc kubenswrapper[4972]: E0228 11:20:00.791071 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:20:00 crc kubenswrapper[4972]: I0228 11:20:00.866695 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537960-pgr9r"] Feb 28 11:20:01 crc kubenswrapper[4972]: I0228 11:20:01.514714 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" event={"ID":"f79393b1-d156-4757-aaff-8b6fb64a57d4","Type":"ContainerStarted","Data":"6d0d5194a55c9441f7c8384d366257a6e1316dc2081a14d0464bbf6e1b7a7421"} Feb 28 11:20:02 crc kubenswrapper[4972]: I0228 11:20:02.530002 4972 generic.go:334] "Generic (PLEG): container finished" podID="f79393b1-d156-4757-aaff-8b6fb64a57d4" containerID="6d6ca20bf89342a288e66a1479887bc7dc27f702242885ca3dbf3d2cae14c8b2" exitCode=0 Feb 28 11:20:02 crc kubenswrapper[4972]: I0228 11:20:02.531426 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" event={"ID":"f79393b1-d156-4757-aaff-8b6fb64a57d4","Type":"ContainerDied","Data":"6d6ca20bf89342a288e66a1479887bc7dc27f702242885ca3dbf3d2cae14c8b2"} Feb 28 11:20:03 crc kubenswrapper[4972]: I0228 11:20:03.964112 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.095367 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mrr\" (UniqueName: \"kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr\") pod \"f79393b1-d156-4757-aaff-8b6fb64a57d4\" (UID: \"f79393b1-d156-4757-aaff-8b6fb64a57d4\") " Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.106949 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr" (OuterVolumeSpecName: "kube-api-access-v7mrr") pod "f79393b1-d156-4757-aaff-8b6fb64a57d4" (UID: "f79393b1-d156-4757-aaff-8b6fb64a57d4"). InnerVolumeSpecName "kube-api-access-v7mrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.198791 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mrr\" (UniqueName: \"kubernetes.io/projected/f79393b1-d156-4757-aaff-8b6fb64a57d4-kube-api-access-v7mrr\") on node \"crc\" DevicePath \"\"" Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.561358 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" event={"ID":"f79393b1-d156-4757-aaff-8b6fb64a57d4","Type":"ContainerDied","Data":"6d0d5194a55c9441f7c8384d366257a6e1316dc2081a14d0464bbf6e1b7a7421"} Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.561427 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d0d5194a55c9441f7c8384d366257a6e1316dc2081a14d0464bbf6e1b7a7421" Feb 28 11:20:04 crc kubenswrapper[4972]: I0228 11:20:04.561946 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537960-pgr9r" Feb 28 11:20:05 crc kubenswrapper[4972]: I0228 11:20:05.051756 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537954-6xn9d"] Feb 28 11:20:05 crc kubenswrapper[4972]: I0228 11:20:05.065657 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537954-6xn9d"] Feb 28 11:20:05 crc kubenswrapper[4972]: I0228 11:20:05.803724 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04322d21-316d-4235-9866-58ecdabedc3c" path="/var/lib/kubelet/pods/04322d21-316d-4235-9866-58ecdabedc3c/volumes" Feb 28 11:20:15 crc kubenswrapper[4972]: I0228 11:20:15.790544 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:20:15 crc kubenswrapper[4972]: E0228 11:20:15.791709 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:20:27 crc kubenswrapper[4972]: I0228 11:20:27.790639 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:20:27 crc kubenswrapper[4972]: E0228 11:20:27.792017 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.766307 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 28 11:20:32 crc kubenswrapper[4972]: E0228 11:20:32.767877 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f79393b1-d156-4757-aaff-8b6fb64a57d4" containerName="oc" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.767903 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="f79393b1-d156-4757-aaff-8b6fb64a57d4" containerName="oc" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.768317 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="f79393b1-d156-4757-aaff-8b6fb64a57d4" containerName="oc" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.769432 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.773804 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.774256 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vgjz8" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.774888 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.776749 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.788076 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.806552 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.806616 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.806898 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908653 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908757 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908828 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908856 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwwg6\" (UniqueName: \"kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908965 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.908998 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.909263 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.909358 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.909526 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.911213 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.911875 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:32 crc kubenswrapper[4972]: I0228 11:20:32.919796 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.012867 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.012983 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.013081 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.013190 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.013237 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwwg6\" (UniqueName: \"kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.013339 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.014221 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.014314 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.014371 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.021344 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.022309 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.045585 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwwg6\" (UniqueName: \"kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.066080 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.116846 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.441229 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 28 11:20:33 crc kubenswrapper[4972]: I0228 11:20:33.939358 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"50684e40-2fa2-4cb3-a176-374ab716698c","Type":"ContainerStarted","Data":"94eddbd462fbbd91895e6276cefcf32288bfeb0bf8d504e85b43b329e8036d2e"} Feb 28 11:20:42 crc kubenswrapper[4972]: I0228 11:20:42.789424 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:20:42 crc kubenswrapper[4972]: E0228 11:20:42.790738 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:20:53 crc kubenswrapper[4972]: I0228 11:20:53.797881 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:20:53 crc kubenswrapper[4972]: E0228 11:20:53.798828 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:21:04 crc kubenswrapper[4972]: I0228 11:21:04.386049 4972 scope.go:117] "RemoveContainer" containerID="647502943ba3db1a14f900a8e6b5275cd466122a8d2b9a7d973f17a032e3fa91" Feb 28 11:21:08 crc kubenswrapper[4972]: I0228 11:21:08.790025 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:21:08 crc kubenswrapper[4972]: E0228 11:21:08.791278 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:21:09 crc kubenswrapper[4972]: E0228 11:21:09.557338 4972 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 28 11:21:09 crc kubenswrapper[4972]: E0228 11:21:09.558183 4972 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jwwg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(50684e40-2fa2-4cb3-a176-374ab716698c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 28 11:21:09 crc kubenswrapper[4972]: E0228 11:21:09.559595 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="50684e40-2fa2-4cb3-a176-374ab716698c" Feb 28 11:21:10 crc kubenswrapper[4972]: E0228 11:21:10.400439 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="50684e40-2fa2-4cb3-a176-374ab716698c" Feb 28 11:21:20 crc kubenswrapper[4972]: I0228 11:21:20.790835 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:21:20 crc kubenswrapper[4972]: E0228 11:21:20.791996 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:21:23 crc kubenswrapper[4972]: I0228 11:21:23.801863 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:21:24 crc kubenswrapper[4972]: I0228 11:21:24.254081 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 28 11:21:25 crc kubenswrapper[4972]: I0228 11:21:25.586881 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"50684e40-2fa2-4cb3-a176-374ab716698c","Type":"ContainerStarted","Data":"5aae7280aa24c06af05541d09a1032228451f1ae547d9db972bc67bce2aa9374"} Feb 28 11:21:25 crc kubenswrapper[4972]: I0228 11:21:25.619352 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.816432296 podStartE2EDuration="54.619326208s" podCreationTimestamp="2026-02-28 11:20:31 +0000 UTC" firstStartedPulling="2026-02-28 11:20:33.446380834 +0000 UTC m=+2690.358388572" lastFinishedPulling="2026-02-28 11:21:24.249274736 +0000 UTC m=+2741.161282484" observedRunningTime="2026-02-28 11:21:25.608393599 +0000 UTC m=+2742.520401347" watchObservedRunningTime="2026-02-28 11:21:25.619326208 +0000 UTC m=+2742.531333946" Feb 28 11:21:35 crc kubenswrapper[4972]: I0228 11:21:35.789506 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:21:35 crc kubenswrapper[4972]: E0228 11:21:35.790757 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:21:46 crc kubenswrapper[4972]: I0228 11:21:46.790141 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:21:46 crc kubenswrapper[4972]: E0228 11:21:46.791034 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:21:59 crc kubenswrapper[4972]: I0228 11:21:59.790444 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:21:59 crc kubenswrapper[4972]: E0228 11:21:59.791309 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.158285 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537962-srlzk"] Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.160630 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.163306 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.163799 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.168062 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.170060 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537962-srlzk"] Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.223551 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p426\" (UniqueName: \"kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426\") pod \"auto-csr-approver-29537962-srlzk\" (UID: \"5a9de5d4-56bc-4f9e-a3ef-57925038c881\") " pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.326007 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p426\" (UniqueName: \"kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426\") pod \"auto-csr-approver-29537962-srlzk\" (UID: \"5a9de5d4-56bc-4f9e-a3ef-57925038c881\") " pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.346820 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p426\" (UniqueName: \"kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426\") pod \"auto-csr-approver-29537962-srlzk\" (UID: \"5a9de5d4-56bc-4f9e-a3ef-57925038c881\") " pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:00 crc kubenswrapper[4972]: I0228 11:22:00.481954 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:01 crc kubenswrapper[4972]: I0228 11:22:01.010561 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537962-srlzk"] Feb 28 11:22:02 crc kubenswrapper[4972]: I0228 11:22:02.021719 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537962-srlzk" event={"ID":"5a9de5d4-56bc-4f9e-a3ef-57925038c881","Type":"ContainerStarted","Data":"900fb21ae1062a93dd290eccf4792927210542258c51c1c7a4604d07bfb32ff8"} Feb 28 11:22:03 crc kubenswrapper[4972]: I0228 11:22:03.034119 4972 generic.go:334] "Generic (PLEG): container finished" podID="5a9de5d4-56bc-4f9e-a3ef-57925038c881" containerID="71efbeda45649382ea33e095ccdfc6a4e2d1755714249e8f3f94e481e4bf563a" exitCode=0 Feb 28 11:22:03 crc kubenswrapper[4972]: I0228 11:22:03.034209 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537962-srlzk" event={"ID":"5a9de5d4-56bc-4f9e-a3ef-57925038c881","Type":"ContainerDied","Data":"71efbeda45649382ea33e095ccdfc6a4e2d1755714249e8f3f94e481e4bf563a"} Feb 28 11:22:04 crc kubenswrapper[4972]: I0228 11:22:04.563156 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:04 crc kubenswrapper[4972]: I0228 11:22:04.640490 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p426\" (UniqueName: \"kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426\") pod \"5a9de5d4-56bc-4f9e-a3ef-57925038c881\" (UID: \"5a9de5d4-56bc-4f9e-a3ef-57925038c881\") " Feb 28 11:22:04 crc kubenswrapper[4972]: I0228 11:22:04.651749 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426" (OuterVolumeSpecName: "kube-api-access-6p426") pod "5a9de5d4-56bc-4f9e-a3ef-57925038c881" (UID: "5a9de5d4-56bc-4f9e-a3ef-57925038c881"). InnerVolumeSpecName "kube-api-access-6p426". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:22:04 crc kubenswrapper[4972]: I0228 11:22:04.743660 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p426\" (UniqueName: \"kubernetes.io/projected/5a9de5d4-56bc-4f9e-a3ef-57925038c881-kube-api-access-6p426\") on node \"crc\" DevicePath \"\"" Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.057319 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537962-srlzk" event={"ID":"5a9de5d4-56bc-4f9e-a3ef-57925038c881","Type":"ContainerDied","Data":"900fb21ae1062a93dd290eccf4792927210542258c51c1c7a4604d07bfb32ff8"} Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.057384 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="900fb21ae1062a93dd290eccf4792927210542258c51c1c7a4604d07bfb32ff8" Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.057513 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537962-srlzk" Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.652990 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537956-zwpgk"] Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.665790 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537956-zwpgk"] Feb 28 11:22:05 crc kubenswrapper[4972]: I0228 11:22:05.802049 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f18cd6e-2519-4ee0-bcf7-9ade01703f7f" path="/var/lib/kubelet/pods/9f18cd6e-2519-4ee0-bcf7-9ade01703f7f/volumes" Feb 28 11:22:09 crc kubenswrapper[4972]: I0228 11:22:09.813695 4972 scope.go:117] "RemoveContainer" containerID="bf80b48e1056cf9f4b6476b37554a153eb6a522ce301d7ac1395479ffc334599" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.663789 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" podUID="510043b9-9bba-49a8-9902-e37f4564bbe8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.767133 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-55f4bf89cb-jm2wm" podUID="510043b9-9bba-49a8-9902-e37f4564bbe8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.887680 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:11 crc kubenswrapper[4972]: E0228 11:22:11.888135 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9de5d4-56bc-4f9e-a3ef-57925038c881" containerName="oc" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.888158 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9de5d4-56bc-4f9e-a3ef-57925038c881" containerName="oc" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.888430 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9de5d4-56bc-4f9e-a3ef-57925038c881" containerName="oc" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.891414 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:11 crc kubenswrapper[4972]: I0228 11:22:11.902634 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.062184 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.062686 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.063035 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjh4\" (UniqueName: \"kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.164948 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjh4\" (UniqueName: \"kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.165039 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.165123 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.165649 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.165929 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.191180 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjh4\" (UniqueName: \"kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4\") pod \"certified-operators-f8hn2\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:12 crc kubenswrapper[4972]: I0228 11:22:12.261971 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:13 crc kubenswrapper[4972]: I0228 11:22:13.516732 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:13 crc kubenswrapper[4972]: W0228 11:22:13.530654 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64ea8fb6_98f1_4cc7_b4ae_284a84eb09c3.slice/crio-3ec3a440917af776e75c30a8963945c4980e57c8dc73dbbcdb19d26e2c3465d4 WatchSource:0}: Error finding container 3ec3a440917af776e75c30a8963945c4980e57c8dc73dbbcdb19d26e2c3465d4: Status 404 returned error can't find the container with id 3ec3a440917af776e75c30a8963945c4980e57c8dc73dbbcdb19d26e2c3465d4 Feb 28 11:22:13 crc kubenswrapper[4972]: I0228 11:22:13.838318 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerStarted","Data":"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6"} Feb 28 11:22:13 crc kubenswrapper[4972]: I0228 11:22:13.838786 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerStarted","Data":"3ec3a440917af776e75c30a8963945c4980e57c8dc73dbbcdb19d26e2c3465d4"} Feb 28 11:22:14 crc kubenswrapper[4972]: I0228 11:22:14.788834 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:22:14 crc kubenswrapper[4972]: E0228 11:22:14.789453 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:22:14 crc kubenswrapper[4972]: I0228 11:22:14.857042 4972 generic.go:334] "Generic (PLEG): container finished" podID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerID="131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6" exitCode=0 Feb 28 11:22:14 crc kubenswrapper[4972]: I0228 11:22:14.857099 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerDied","Data":"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6"} Feb 28 11:22:14 crc kubenswrapper[4972]: I0228 11:22:14.857131 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerStarted","Data":"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e"} Feb 28 11:22:15 crc kubenswrapper[4972]: I0228 11:22:15.875514 4972 generic.go:334] "Generic (PLEG): container finished" podID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerID="6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e" exitCode=0 Feb 28 11:22:15 crc kubenswrapper[4972]: I0228 11:22:15.875572 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerDied","Data":"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e"} Feb 28 11:22:16 crc kubenswrapper[4972]: I0228 11:22:16.888326 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerStarted","Data":"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c"} Feb 28 11:22:16 crc kubenswrapper[4972]: I0228 11:22:16.912244 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f8hn2" podStartSLOduration=3.475795045 podStartE2EDuration="5.91222423s" podCreationTimestamp="2026-02-28 11:22:11 +0000 UTC" firstStartedPulling="2026-02-28 11:22:13.844722807 +0000 UTC m=+2790.756730585" lastFinishedPulling="2026-02-28 11:22:16.281152032 +0000 UTC m=+2793.193159770" observedRunningTime="2026-02-28 11:22:16.90458901 +0000 UTC m=+2793.816596748" watchObservedRunningTime="2026-02-28 11:22:16.91222423 +0000 UTC m=+2793.824231968" Feb 28 11:22:22 crc kubenswrapper[4972]: I0228 11:22:22.262543 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:22 crc kubenswrapper[4972]: I0228 11:22:22.263137 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:22 crc kubenswrapper[4972]: I0228 11:22:22.346105 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:23 crc kubenswrapper[4972]: I0228 11:22:23.017331 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:23 crc kubenswrapper[4972]: I0228 11:22:23.107754 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:24 crc kubenswrapper[4972]: I0228 11:22:24.972398 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f8hn2" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="registry-server" containerID="cri-o://d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c" gracePeriod=2 Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.516812 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.582364 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjh4\" (UniqueName: \"kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4\") pod \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.582505 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities\") pod \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.582560 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content\") pod \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\" (UID: \"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3\") " Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.586488 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities" (OuterVolumeSpecName: "utilities") pod "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" (UID: "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.615942 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4" (OuterVolumeSpecName: "kube-api-access-vkjh4") pod "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" (UID: "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3"). InnerVolumeSpecName "kube-api-access-vkjh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.685142 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.685192 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkjh4\" (UniqueName: \"kubernetes.io/projected/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-kube-api-access-vkjh4\") on node \"crc\" DevicePath \"\"" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.996206 4972 generic.go:334] "Generic (PLEG): container finished" podID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerID="d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c" exitCode=0 Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.996669 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerDied","Data":"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c"} Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.996705 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8hn2" event={"ID":"64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3","Type":"ContainerDied","Data":"3ec3a440917af776e75c30a8963945c4980e57c8dc73dbbcdb19d26e2c3465d4"} Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.996730 4972 scope.go:117] "RemoveContainer" containerID="d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c" Feb 28 11:22:25 crc kubenswrapper[4972]: I0228 11:22:25.996913 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8hn2" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.028638 4972 scope.go:117] "RemoveContainer" containerID="6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.073288 4972 scope.go:117] "RemoveContainer" containerID="131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.144794 4972 scope.go:117] "RemoveContainer" containerID="d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c" Feb 28 11:22:26 crc kubenswrapper[4972]: E0228 11:22:26.146094 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c\": container with ID starting with d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c not found: ID does not exist" containerID="d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.146149 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c"} err="failed to get container status \"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c\": rpc error: code = NotFound desc = could not find container \"d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c\": container with ID starting with d6863230448a217c462a8c7f387e153af01982a1a1404e1d47df77814b35186c not found: ID does not exist" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.146184 4972 scope.go:117] "RemoveContainer" containerID="6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e" Feb 28 11:22:26 crc kubenswrapper[4972]: E0228 11:22:26.146773 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e\": container with ID starting with 6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e not found: ID does not exist" containerID="6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.146902 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e"} err="failed to get container status \"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e\": rpc error: code = NotFound desc = could not find container \"6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e\": container with ID starting with 6b5bdb80c2399d62aaa449dbbd1bb638abaa832668316f29867e9ea7ca5aed1e not found: ID does not exist" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.147040 4972 scope.go:117] "RemoveContainer" containerID="131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6" Feb 28 11:22:26 crc kubenswrapper[4972]: E0228 11:22:26.147808 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6\": container with ID starting with 131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6 not found: ID does not exist" containerID="131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.147902 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6"} err="failed to get container status \"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6\": rpc error: code = NotFound desc = could not find container \"131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6\": container with ID starting with 131f779d91ef36cb8ece6c08126fe0e76c0dabd7551fddc1e03e0d6bd3f739c6 not found: ID does not exist" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.186685 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" (UID: "64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.198517 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.354284 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:26 crc kubenswrapper[4972]: I0228 11:22:26.363320 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f8hn2"] Feb 28 11:22:27 crc kubenswrapper[4972]: I0228 11:22:27.789979 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:22:27 crc kubenswrapper[4972]: E0228 11:22:27.790536 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:22:27 crc kubenswrapper[4972]: I0228 11:22:27.799806 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" path="/var/lib/kubelet/pods/64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3/volumes" Feb 28 11:22:39 crc kubenswrapper[4972]: I0228 11:22:39.789640 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:22:40 crc kubenswrapper[4972]: I0228 11:22:40.149104 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024"} Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.167933 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537964-2cbzx"] Feb 28 11:24:00 crc kubenswrapper[4972]: E0228 11:24:00.169407 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="registry-server" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.169438 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="registry-server" Feb 28 11:24:00 crc kubenswrapper[4972]: E0228 11:24:00.169523 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="extract-utilities" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.169543 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="extract-utilities" Feb 28 11:24:00 crc kubenswrapper[4972]: E0228 11:24:00.169570 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="extract-content" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.169587 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="extract-content" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.170000 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ea8fb6-98f1-4cc7-b4ae-284a84eb09c3" containerName="registry-server" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.171207 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.173156 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.176931 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.177187 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.180631 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537964-2cbzx"] Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.338979 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q298w\" (UniqueName: \"kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w\") pod \"auto-csr-approver-29537964-2cbzx\" (UID: \"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a\") " pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.441006 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q298w\" (UniqueName: \"kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w\") pod \"auto-csr-approver-29537964-2cbzx\" (UID: \"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a\") " pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.478771 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q298w\" (UniqueName: \"kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w\") pod \"auto-csr-approver-29537964-2cbzx\" (UID: \"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a\") " pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:00 crc kubenswrapper[4972]: I0228 11:24:00.497766 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:01 crc kubenswrapper[4972]: I0228 11:24:01.047045 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537964-2cbzx"] Feb 28 11:24:01 crc kubenswrapper[4972]: I0228 11:24:01.211529 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" event={"ID":"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a","Type":"ContainerStarted","Data":"dc2b01998ee0a8662c0ded09d96e091f1d91746706e34a3ac9e140407596fb4f"} Feb 28 11:24:03 crc kubenswrapper[4972]: I0228 11:24:03.242124 4972 generic.go:334] "Generic (PLEG): container finished" podID="7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" containerID="2f0f99b14f7575771ab0150c659d1ef9b508ada9404924c7544bbc509a64cf08" exitCode=0 Feb 28 11:24:03 crc kubenswrapper[4972]: I0228 11:24:03.242202 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" event={"ID":"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a","Type":"ContainerDied","Data":"2f0f99b14f7575771ab0150c659d1ef9b508ada9404924c7544bbc509a64cf08"} Feb 28 11:24:04 crc kubenswrapper[4972]: I0228 11:24:04.735480 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:04 crc kubenswrapper[4972]: I0228 11:24:04.846425 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q298w\" (UniqueName: \"kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w\") pod \"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a\" (UID: \"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a\") " Feb 28 11:24:04 crc kubenswrapper[4972]: I0228 11:24:04.852607 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w" (OuterVolumeSpecName: "kube-api-access-q298w") pod "7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" (UID: "7fe5ae58-c5f0-4540-aa4a-191bbde11f2a"). InnerVolumeSpecName "kube-api-access-q298w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:24:04 crc kubenswrapper[4972]: I0228 11:24:04.949166 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q298w\" (UniqueName: \"kubernetes.io/projected/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a-kube-api-access-q298w\") on node \"crc\" DevicePath \"\"" Feb 28 11:24:05 crc kubenswrapper[4972]: I0228 11:24:05.269137 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" event={"ID":"7fe5ae58-c5f0-4540-aa4a-191bbde11f2a","Type":"ContainerDied","Data":"dc2b01998ee0a8662c0ded09d96e091f1d91746706e34a3ac9e140407596fb4f"} Feb 28 11:24:05 crc kubenswrapper[4972]: I0228 11:24:05.269185 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc2b01998ee0a8662c0ded09d96e091f1d91746706e34a3ac9e140407596fb4f" Feb 28 11:24:05 crc kubenswrapper[4972]: I0228 11:24:05.269201 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537964-2cbzx" Feb 28 11:24:05 crc kubenswrapper[4972]: I0228 11:24:05.812819 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537958-thnxn"] Feb 28 11:24:05 crc kubenswrapper[4972]: I0228 11:24:05.830685 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537958-thnxn"] Feb 28 11:24:07 crc kubenswrapper[4972]: I0228 11:24:07.812585 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ccb883b-1ed6-4dfc-a564-82a65b472b1f" path="/var/lib/kubelet/pods/2ccb883b-1ed6-4dfc-a564-82a65b472b1f/volumes" Feb 28 11:24:09 crc kubenswrapper[4972]: I0228 11:24:09.980085 4972 scope.go:117] "RemoveContainer" containerID="d0649da349332a55de361d852164f2550873209b3a8e4a4f5b35b13f312b6d34" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.489499 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:24:52 crc kubenswrapper[4972]: E0228 11:24:52.490818 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" containerName="oc" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.490840 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" containerName="oc" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.491386 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" containerName="oc" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.494678 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.513805 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.682491 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl9n7\" (UniqueName: \"kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.682587 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.682749 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.784581 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.784726 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl9n7\" (UniqueName: \"kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.784795 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.785475 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.785549 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.814237 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl9n7\" (UniqueName: \"kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7\") pod \"redhat-marketplace-xrmc8\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:52 crc kubenswrapper[4972]: I0228 11:24:52.829029 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:24:53 crc kubenswrapper[4972]: I0228 11:24:53.346124 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:24:53 crc kubenswrapper[4972]: I0228 11:24:53.891608 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad01333f-0298-4844-9679-03b77d8c8098" containerID="c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c" exitCode=0 Feb 28 11:24:53 crc kubenswrapper[4972]: I0228 11:24:53.891701 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerDied","Data":"c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c"} Feb 28 11:24:53 crc kubenswrapper[4972]: I0228 11:24:53.892150 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerStarted","Data":"84763369590ab44c380f3f4de6b202436e4e56678204b25bd4cc47845990ab3c"} Feb 28 11:24:54 crc kubenswrapper[4972]: I0228 11:24:54.915607 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerStarted","Data":"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78"} Feb 28 11:24:56 crc kubenswrapper[4972]: I0228 11:24:56.939796 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad01333f-0298-4844-9679-03b77d8c8098" containerID="a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78" exitCode=0 Feb 28 11:24:56 crc kubenswrapper[4972]: I0228 11:24:56.939928 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerDied","Data":"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78"} Feb 28 11:24:57 crc kubenswrapper[4972]: I0228 11:24:57.955960 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerStarted","Data":"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b"} Feb 28 11:24:57 crc kubenswrapper[4972]: I0228 11:24:57.994200 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xrmc8" podStartSLOduration=2.517638583 podStartE2EDuration="5.994127879s" podCreationTimestamp="2026-02-28 11:24:52 +0000 UTC" firstStartedPulling="2026-02-28 11:24:53.894769653 +0000 UTC m=+2950.806777431" lastFinishedPulling="2026-02-28 11:24:57.371258969 +0000 UTC m=+2954.283266727" observedRunningTime="2026-02-28 11:24:57.981147387 +0000 UTC m=+2954.893155155" watchObservedRunningTime="2026-02-28 11:24:57.994127879 +0000 UTC m=+2954.906135657" Feb 28 11:24:58 crc kubenswrapper[4972]: I0228 11:24:58.890392 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:24:58 crc kubenswrapper[4972]: I0228 11:24:58.890508 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:25:02 crc kubenswrapper[4972]: I0228 11:25:02.829949 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:02 crc kubenswrapper[4972]: I0228 11:25:02.830724 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:02 crc kubenswrapper[4972]: I0228 11:25:02.903288 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:03 crc kubenswrapper[4972]: I0228 11:25:03.068630 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:03 crc kubenswrapper[4972]: I0228 11:25:03.160213 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.029133 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xrmc8" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="registry-server" containerID="cri-o://a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b" gracePeriod=2 Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.536524 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.727253 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl9n7\" (UniqueName: \"kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7\") pod \"ad01333f-0298-4844-9679-03b77d8c8098\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.727676 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content\") pod \"ad01333f-0298-4844-9679-03b77d8c8098\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.727746 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities\") pod \"ad01333f-0298-4844-9679-03b77d8c8098\" (UID: \"ad01333f-0298-4844-9679-03b77d8c8098\") " Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.728821 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities" (OuterVolumeSpecName: "utilities") pod "ad01333f-0298-4844-9679-03b77d8c8098" (UID: "ad01333f-0298-4844-9679-03b77d8c8098"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.735237 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7" (OuterVolumeSpecName: "kube-api-access-kl9n7") pod "ad01333f-0298-4844-9679-03b77d8c8098" (UID: "ad01333f-0298-4844-9679-03b77d8c8098"). InnerVolumeSpecName "kube-api-access-kl9n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.763818 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad01333f-0298-4844-9679-03b77d8c8098" (UID: "ad01333f-0298-4844-9679-03b77d8c8098"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.830040 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl9n7\" (UniqueName: \"kubernetes.io/projected/ad01333f-0298-4844-9679-03b77d8c8098-kube-api-access-kl9n7\") on node \"crc\" DevicePath \"\"" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.830287 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:25:05 crc kubenswrapper[4972]: I0228 11:25:05.831144 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad01333f-0298-4844-9679-03b77d8c8098-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.046687 4972 generic.go:334] "Generic (PLEG): container finished" podID="ad01333f-0298-4844-9679-03b77d8c8098" containerID="a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b" exitCode=0 Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.046803 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerDied","Data":"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b"} Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.046905 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xrmc8" event={"ID":"ad01333f-0298-4844-9679-03b77d8c8098","Type":"ContainerDied","Data":"84763369590ab44c380f3f4de6b202436e4e56678204b25bd4cc47845990ab3c"} Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.046943 4972 scope.go:117] "RemoveContainer" containerID="a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.046940 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xrmc8" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.094524 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.095708 4972 scope.go:117] "RemoveContainer" containerID="a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.111658 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xrmc8"] Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.145958 4972 scope.go:117] "RemoveContainer" containerID="c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.206988 4972 scope.go:117] "RemoveContainer" containerID="a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b" Feb 28 11:25:06 crc kubenswrapper[4972]: E0228 11:25:06.207890 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b\": container with ID starting with a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b not found: ID does not exist" containerID="a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.207954 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b"} err="failed to get container status \"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b\": rpc error: code = NotFound desc = could not find container \"a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b\": container with ID starting with a5625980ca5ad46b96006958dc5c99e0a4cd5ea6d8f32a96e27a1c713a49f88b not found: ID does not exist" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.207996 4972 scope.go:117] "RemoveContainer" containerID="a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78" Feb 28 11:25:06 crc kubenswrapper[4972]: E0228 11:25:06.208638 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78\": container with ID starting with a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78 not found: ID does not exist" containerID="a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.208696 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78"} err="failed to get container status \"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78\": rpc error: code = NotFound desc = could not find container \"a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78\": container with ID starting with a1f691c4189729875fab399e46e5115098cca0fa13cecfc14568a6c1e47c7b78 not found: ID does not exist" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.208734 4972 scope.go:117] "RemoveContainer" containerID="c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c" Feb 28 11:25:06 crc kubenswrapper[4972]: E0228 11:25:06.209859 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c\": container with ID starting with c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c not found: ID does not exist" containerID="c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c" Feb 28 11:25:06 crc kubenswrapper[4972]: I0228 11:25:06.209909 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c"} err="failed to get container status \"c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c\": rpc error: code = NotFound desc = could not find container \"c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c\": container with ID starting with c3c8e360272d4352c7c24ce09edf75b6f794c8dc1a57cd5f04c442c85870147c not found: ID does not exist" Feb 28 11:25:07 crc kubenswrapper[4972]: I0228 11:25:07.809367 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad01333f-0298-4844-9679-03b77d8c8098" path="/var/lib/kubelet/pods/ad01333f-0298-4844-9679-03b77d8c8098/volumes" Feb 28 11:25:28 crc kubenswrapper[4972]: I0228 11:25:28.891126 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:25:28 crc kubenswrapper[4972]: I0228 11:25:28.891874 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:25:58 crc kubenswrapper[4972]: I0228 11:25:58.890997 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:25:58 crc kubenswrapper[4972]: I0228 11:25:58.892142 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:25:58 crc kubenswrapper[4972]: I0228 11:25:58.892224 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:25:58 crc kubenswrapper[4972]: I0228 11:25:58.893644 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:25:58 crc kubenswrapper[4972]: I0228 11:25:58.893730 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024" gracePeriod=600 Feb 28 11:25:59 crc kubenswrapper[4972]: I0228 11:25:59.766174 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024" exitCode=0 Feb 28 11:25:59 crc kubenswrapper[4972]: I0228 11:25:59.766320 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024"} Feb 28 11:25:59 crc kubenswrapper[4972]: I0228 11:25:59.766921 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e"} Feb 28 11:25:59 crc kubenswrapper[4972]: I0228 11:25:59.766949 4972 scope.go:117] "RemoveContainer" containerID="8d6cabd5406eb303defade38bc81a217fb4ed2d1fd22a1a9f81e12e7d89664a0" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.142922 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537966-2bwd2"] Feb 28 11:26:00 crc kubenswrapper[4972]: E0228 11:26:00.143567 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="registry-server" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.143580 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="registry-server" Feb 28 11:26:00 crc kubenswrapper[4972]: E0228 11:26:00.143612 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="extract-content" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.143618 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="extract-content" Feb 28 11:26:00 crc kubenswrapper[4972]: E0228 11:26:00.143632 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="extract-utilities" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.143640 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="extract-utilities" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.143812 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad01333f-0298-4844-9679-03b77d8c8098" containerName="registry-server" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.144390 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.146644 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.147270 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.147719 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.169805 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6tqr\" (UniqueName: \"kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr\") pod \"auto-csr-approver-29537966-2bwd2\" (UID: \"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c\") " pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.176339 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537966-2bwd2"] Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.271452 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6tqr\" (UniqueName: \"kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr\") pod \"auto-csr-approver-29537966-2bwd2\" (UID: \"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c\") " pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.309190 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6tqr\" (UniqueName: \"kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr\") pod \"auto-csr-approver-29537966-2bwd2\" (UID: \"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c\") " pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.463139 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:00 crc kubenswrapper[4972]: I0228 11:26:00.995495 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537966-2bwd2"] Feb 28 11:26:01 crc kubenswrapper[4972]: I0228 11:26:01.818372 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" event={"ID":"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c","Type":"ContainerStarted","Data":"ce03c13f09264cf5b4cdd5804973d54075ad3fb3fc3aaa71f5d692a2f7403d6d"} Feb 28 11:26:02 crc kubenswrapper[4972]: I0228 11:26:02.819238 4972 generic.go:334] "Generic (PLEG): container finished" podID="357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" containerID="a06463251c2f039b055f84d90a36ca15275586d0c8a13ab28c58358cd09932d4" exitCode=0 Feb 28 11:26:02 crc kubenswrapper[4972]: I0228 11:26:02.819328 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" event={"ID":"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c","Type":"ContainerDied","Data":"a06463251c2f039b055f84d90a36ca15275586d0c8a13ab28c58358cd09932d4"} Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.316710 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.362258 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6tqr\" (UniqueName: \"kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr\") pod \"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c\" (UID: \"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c\") " Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.371845 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr" (OuterVolumeSpecName: "kube-api-access-q6tqr") pod "357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" (UID: "357dbfdb-c40c-49b6-a1d2-ffdca782fe7c"). InnerVolumeSpecName "kube-api-access-q6tqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.465935 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6tqr\" (UniqueName: \"kubernetes.io/projected/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c-kube-api-access-q6tqr\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.848119 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" event={"ID":"357dbfdb-c40c-49b6-a1d2-ffdca782fe7c","Type":"ContainerDied","Data":"ce03c13f09264cf5b4cdd5804973d54075ad3fb3fc3aaa71f5d692a2f7403d6d"} Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.848194 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce03c13f09264cf5b4cdd5804973d54075ad3fb3fc3aaa71f5d692a2f7403d6d" Feb 28 11:26:04 crc kubenswrapper[4972]: I0228 11:26:04.848250 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537966-2bwd2" Feb 28 11:26:05 crc kubenswrapper[4972]: I0228 11:26:05.430523 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537960-pgr9r"] Feb 28 11:26:05 crc kubenswrapper[4972]: I0228 11:26:05.440542 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537960-pgr9r"] Feb 28 11:26:05 crc kubenswrapper[4972]: I0228 11:26:05.806309 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f79393b1-d156-4757-aaff-8b6fb64a57d4" path="/var/lib/kubelet/pods/f79393b1-d156-4757-aaff-8b6fb64a57d4/volumes" Feb 28 11:26:10 crc kubenswrapper[4972]: I0228 11:26:10.177927 4972 scope.go:117] "RemoveContainer" containerID="6d6ca20bf89342a288e66a1479887bc7dc27f702242885ca3dbf3d2cae14c8b2" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.150796 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:32 crc kubenswrapper[4972]: E0228 11:26:32.151997 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" containerName="oc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.152019 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" containerName="oc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.152217 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" containerName="oc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.153950 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.205366 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.232777 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.232982 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.233229 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jps9t\" (UniqueName: \"kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.335168 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.335303 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jps9t\" (UniqueName: \"kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.335485 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.336077 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.336110 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.361435 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jps9t\" (UniqueName: \"kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t\") pod \"community-operators-p4fdc\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:32 crc kubenswrapper[4972]: I0228 11:26:32.492509 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:33 crc kubenswrapper[4972]: I0228 11:26:33.082238 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:33 crc kubenswrapper[4972]: I0228 11:26:33.202198 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerStarted","Data":"26081b027dd86c5e21d7e2747798fd775c5ed07305b7e78b8b227727b1f1536f"} Feb 28 11:26:34 crc kubenswrapper[4972]: I0228 11:26:34.215093 4972 generic.go:334] "Generic (PLEG): container finished" podID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerID="9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8" exitCode=0 Feb 28 11:26:34 crc kubenswrapper[4972]: I0228 11:26:34.215188 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerDied","Data":"9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8"} Feb 28 11:26:34 crc kubenswrapper[4972]: I0228 11:26:34.218616 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.148920 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.153283 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.171270 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.233200 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.233324 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.233965 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bznmw\" (UniqueName: \"kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.335898 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bznmw\" (UniqueName: \"kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.335956 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.336003 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.336506 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.336572 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.361620 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bznmw\" (UniqueName: \"kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw\") pod \"redhat-operators-jxpqq\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.492338 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:35 crc kubenswrapper[4972]: I0228 11:26:35.962654 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:36 crc kubenswrapper[4972]: I0228 11:26:36.268441 4972 generic.go:334] "Generic (PLEG): container finished" podID="988221f4-e930-4044-8144-c293a3ad3a59" containerID="5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095" exitCode=0 Feb 28 11:26:36 crc kubenswrapper[4972]: I0228 11:26:36.268545 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerDied","Data":"5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095"} Feb 28 11:26:36 crc kubenswrapper[4972]: I0228 11:26:36.268577 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerStarted","Data":"e356dc9dee086095b27df3028e4355346b4bc263067ba2eabdee6d3c102dca59"} Feb 28 11:26:36 crc kubenswrapper[4972]: I0228 11:26:36.291487 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerStarted","Data":"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf"} Feb 28 11:26:37 crc kubenswrapper[4972]: I0228 11:26:37.306996 4972 generic.go:334] "Generic (PLEG): container finished" podID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerID="27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf" exitCode=0 Feb 28 11:26:37 crc kubenswrapper[4972]: I0228 11:26:37.307057 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerDied","Data":"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf"} Feb 28 11:26:37 crc kubenswrapper[4972]: I0228 11:26:37.317927 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerStarted","Data":"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4"} Feb 28 11:26:38 crc kubenswrapper[4972]: I0228 11:26:38.331832 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerStarted","Data":"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab"} Feb 28 11:26:38 crc kubenswrapper[4972]: I0228 11:26:38.361579 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4fdc" podStartSLOduration=2.750808238 podStartE2EDuration="6.361554009s" podCreationTimestamp="2026-02-28 11:26:32 +0000 UTC" firstStartedPulling="2026-02-28 11:26:34.217826004 +0000 UTC m=+3051.129833782" lastFinishedPulling="2026-02-28 11:26:37.828571785 +0000 UTC m=+3054.740579553" observedRunningTime="2026-02-28 11:26:38.355881336 +0000 UTC m=+3055.267889074" watchObservedRunningTime="2026-02-28 11:26:38.361554009 +0000 UTC m=+3055.273561787" Feb 28 11:26:42 crc kubenswrapper[4972]: I0228 11:26:42.380961 4972 generic.go:334] "Generic (PLEG): container finished" podID="988221f4-e930-4044-8144-c293a3ad3a59" containerID="58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4" exitCode=0 Feb 28 11:26:42 crc kubenswrapper[4972]: I0228 11:26:42.381049 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerDied","Data":"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4"} Feb 28 11:26:42 crc kubenswrapper[4972]: I0228 11:26:42.493169 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:42 crc kubenswrapper[4972]: I0228 11:26:42.493266 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:43 crc kubenswrapper[4972]: I0228 11:26:43.401684 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerStarted","Data":"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a"} Feb 28 11:26:43 crc kubenswrapper[4972]: I0228 11:26:43.429959 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jxpqq" podStartSLOduration=1.9012963410000001 podStartE2EDuration="8.429939518s" podCreationTimestamp="2026-02-28 11:26:35 +0000 UTC" firstStartedPulling="2026-02-28 11:26:36.270869207 +0000 UTC m=+3053.182876955" lastFinishedPulling="2026-02-28 11:26:42.799512404 +0000 UTC m=+3059.711520132" observedRunningTime="2026-02-28 11:26:43.426607787 +0000 UTC m=+3060.338615555" watchObservedRunningTime="2026-02-28 11:26:43.429939518 +0000 UTC m=+3060.341947266" Feb 28 11:26:43 crc kubenswrapper[4972]: I0228 11:26:43.572928 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-p4fdc" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="registry-server" probeResult="failure" output=< Feb 28 11:26:43 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:26:43 crc kubenswrapper[4972]: > Feb 28 11:26:45 crc kubenswrapper[4972]: I0228 11:26:45.493152 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:45 crc kubenswrapper[4972]: I0228 11:26:45.493878 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:46 crc kubenswrapper[4972]: I0228 11:26:46.562418 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jxpqq" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="registry-server" probeResult="failure" output=< Feb 28 11:26:46 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:26:46 crc kubenswrapper[4972]: > Feb 28 11:26:52 crc kubenswrapper[4972]: I0228 11:26:52.577548 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:52 crc kubenswrapper[4972]: I0228 11:26:52.652290 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:52 crc kubenswrapper[4972]: I0228 11:26:52.826982 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:54 crc kubenswrapper[4972]: I0228 11:26:54.561195 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4fdc" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="registry-server" containerID="cri-o://06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab" gracePeriod=2 Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.195954 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.288111 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities\") pod \"70e8e523-5550-489b-ae9f-190bfeef78e2\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.288254 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content\") pod \"70e8e523-5550-489b-ae9f-190bfeef78e2\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.288437 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jps9t\" (UniqueName: \"kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t\") pod \"70e8e523-5550-489b-ae9f-190bfeef78e2\" (UID: \"70e8e523-5550-489b-ae9f-190bfeef78e2\") " Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.291266 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities" (OuterVolumeSpecName: "utilities") pod "70e8e523-5550-489b-ae9f-190bfeef78e2" (UID: "70e8e523-5550-489b-ae9f-190bfeef78e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.298005 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t" (OuterVolumeSpecName: "kube-api-access-jps9t") pod "70e8e523-5550-489b-ae9f-190bfeef78e2" (UID: "70e8e523-5550-489b-ae9f-190bfeef78e2"). InnerVolumeSpecName "kube-api-access-jps9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.356525 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70e8e523-5550-489b-ae9f-190bfeef78e2" (UID: "70e8e523-5550-489b-ae9f-190bfeef78e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.391340 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jps9t\" (UniqueName: \"kubernetes.io/projected/70e8e523-5550-489b-ae9f-190bfeef78e2-kube-api-access-jps9t\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.391369 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.391378 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70e8e523-5550-489b-ae9f-190bfeef78e2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.574175 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.579029 4972 generic.go:334] "Generic (PLEG): container finished" podID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerID="06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab" exitCode=0 Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.579094 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4fdc" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.579090 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerDied","Data":"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab"} Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.580787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4fdc" event={"ID":"70e8e523-5550-489b-ae9f-190bfeef78e2","Type":"ContainerDied","Data":"26081b027dd86c5e21d7e2747798fd775c5ed07305b7e78b8b227727b1f1536f"} Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.580848 4972 scope.go:117] "RemoveContainer" containerID="06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.635778 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.639902 4972 scope.go:117] "RemoveContainer" containerID="27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.642028 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.647342 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4fdc"] Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.675544 4972 scope.go:117] "RemoveContainer" containerID="9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.721128 4972 scope.go:117] "RemoveContainer" containerID="06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab" Feb 28 11:26:55 crc kubenswrapper[4972]: E0228 11:26:55.721830 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab\": container with ID starting with 06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab not found: ID does not exist" containerID="06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.721886 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab"} err="failed to get container status \"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab\": rpc error: code = NotFound desc = could not find container \"06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab\": container with ID starting with 06412bda196a689f9128b8ff8e0f63f53007659c1a4833c00c548b8c06a7fbab not found: ID does not exist" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.721921 4972 scope.go:117] "RemoveContainer" containerID="27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf" Feb 28 11:26:55 crc kubenswrapper[4972]: E0228 11:26:55.722401 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf\": container with ID starting with 27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf not found: ID does not exist" containerID="27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.722430 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf"} err="failed to get container status \"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf\": rpc error: code = NotFound desc = could not find container \"27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf\": container with ID starting with 27a4f53864a2250de084b09f0afe884270b592d0940efe0493d540b9caa92abf not found: ID does not exist" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.722446 4972 scope.go:117] "RemoveContainer" containerID="9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8" Feb 28 11:26:55 crc kubenswrapper[4972]: E0228 11:26:55.722849 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8\": container with ID starting with 9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8 not found: ID does not exist" containerID="9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.722913 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8"} err="failed to get container status \"9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8\": rpc error: code = NotFound desc = could not find container \"9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8\": container with ID starting with 9cc56d510117986912437643560fe308d1ff303bd47fe9f97b2107b4d1cef9f8 not found: ID does not exist" Feb 28 11:26:55 crc kubenswrapper[4972]: I0228 11:26:55.804693 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" path="/var/lib/kubelet/pods/70e8e523-5550-489b-ae9f-190bfeef78e2/volumes" Feb 28 11:26:57 crc kubenswrapper[4972]: I0228 11:26:57.835270 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:57 crc kubenswrapper[4972]: I0228 11:26:57.836155 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jxpqq" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="registry-server" containerID="cri-o://f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a" gracePeriod=2 Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.475494 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.569978 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities\") pod \"988221f4-e930-4044-8144-c293a3ad3a59\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.570611 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content\") pod \"988221f4-e930-4044-8144-c293a3ad3a59\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.570763 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bznmw\" (UniqueName: \"kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw\") pod \"988221f4-e930-4044-8144-c293a3ad3a59\" (UID: \"988221f4-e930-4044-8144-c293a3ad3a59\") " Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.571148 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities" (OuterVolumeSpecName: "utilities") pod "988221f4-e930-4044-8144-c293a3ad3a59" (UID: "988221f4-e930-4044-8144-c293a3ad3a59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.571617 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.585868 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw" (OuterVolumeSpecName: "kube-api-access-bznmw") pod "988221f4-e930-4044-8144-c293a3ad3a59" (UID: "988221f4-e930-4044-8144-c293a3ad3a59"). InnerVolumeSpecName "kube-api-access-bznmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.634060 4972 generic.go:334] "Generic (PLEG): container finished" podID="988221f4-e930-4044-8144-c293a3ad3a59" containerID="f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a" exitCode=0 Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.634122 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerDied","Data":"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a"} Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.634167 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jxpqq" event={"ID":"988221f4-e930-4044-8144-c293a3ad3a59","Type":"ContainerDied","Data":"e356dc9dee086095b27df3028e4355346b4bc263067ba2eabdee6d3c102dca59"} Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.634197 4972 scope.go:117] "RemoveContainer" containerID="f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.634397 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jxpqq" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.670184 4972 scope.go:117] "RemoveContainer" containerID="58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.674398 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bznmw\" (UniqueName: \"kubernetes.io/projected/988221f4-e930-4044-8144-c293a3ad3a59-kube-api-access-bznmw\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.714109 4972 scope.go:117] "RemoveContainer" containerID="5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.720406 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "988221f4-e930-4044-8144-c293a3ad3a59" (UID: "988221f4-e930-4044-8144-c293a3ad3a59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.755345 4972 scope.go:117] "RemoveContainer" containerID="f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a" Feb 28 11:26:58 crc kubenswrapper[4972]: E0228 11:26:58.756958 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a\": container with ID starting with f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a not found: ID does not exist" containerID="f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.757069 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a"} err="failed to get container status \"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a\": rpc error: code = NotFound desc = could not find container \"f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a\": container with ID starting with f199de4994fcef724c816019c5448caf70316d316cbcd4fa136c02c1c92f528a not found: ID does not exist" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.757161 4972 scope.go:117] "RemoveContainer" containerID="58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4" Feb 28 11:26:58 crc kubenswrapper[4972]: E0228 11:26:58.757796 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4\": container with ID starting with 58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4 not found: ID does not exist" containerID="58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.757839 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4"} err="failed to get container status \"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4\": rpc error: code = NotFound desc = could not find container \"58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4\": container with ID starting with 58944d8e2bf3d039586e8cffa58aaf54438019f9786585b43e4cc09727d09bc4 not found: ID does not exist" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.757866 4972 scope.go:117] "RemoveContainer" containerID="5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095" Feb 28 11:26:58 crc kubenswrapper[4972]: E0228 11:26:58.758694 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095\": container with ID starting with 5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095 not found: ID does not exist" containerID="5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.758748 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095"} err="failed to get container status \"5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095\": rpc error: code = NotFound desc = could not find container \"5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095\": container with ID starting with 5114b90862e7797f95c50aa5e67527e9315458d1eafb5412c920745e63abb095 not found: ID does not exist" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.776423 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/988221f4-e930-4044-8144-c293a3ad3a59-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.982532 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:58 crc kubenswrapper[4972]: I0228 11:26:58.993651 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jxpqq"] Feb 28 11:26:59 crc kubenswrapper[4972]: I0228 11:26:59.816411 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="988221f4-e930-4044-8144-c293a3ad3a59" path="/var/lib/kubelet/pods/988221f4-e930-4044-8144-c293a3ad3a59/volumes" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.156108 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537968-l8pzf"] Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157246 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="extract-utilities" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157264 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="extract-utilities" Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157287 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="extract-content" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157295 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="extract-content" Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157305 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="extract-content" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157312 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="extract-content" Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157322 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157328 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157349 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="extract-utilities" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157356 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="extract-utilities" Feb 28 11:28:00 crc kubenswrapper[4972]: E0228 11:28:00.157381 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157389 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157622 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="988221f4-e930-4044-8144-c293a3ad3a59" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.157652 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="70e8e523-5550-489b-ae9f-190bfeef78e2" containerName="registry-server" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.158464 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.161666 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.161978 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.162229 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.175174 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9h89\" (UniqueName: \"kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89\") pod \"auto-csr-approver-29537968-l8pzf\" (UID: \"a30a4c89-8677-49ac-8709-a63c48017fd0\") " pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.176910 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537968-l8pzf"] Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.276746 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9h89\" (UniqueName: \"kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89\") pod \"auto-csr-approver-29537968-l8pzf\" (UID: \"a30a4c89-8677-49ac-8709-a63c48017fd0\") " pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.310779 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9h89\" (UniqueName: \"kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89\") pod \"auto-csr-approver-29537968-l8pzf\" (UID: \"a30a4c89-8677-49ac-8709-a63c48017fd0\") " pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:00 crc kubenswrapper[4972]: I0228 11:28:00.480380 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:01 crc kubenswrapper[4972]: I0228 11:28:01.027444 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537968-l8pzf"] Feb 28 11:28:01 crc kubenswrapper[4972]: I0228 11:28:01.609062 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" event={"ID":"a30a4c89-8677-49ac-8709-a63c48017fd0","Type":"ContainerStarted","Data":"d5850ee099ec82fd0fb5020b789c42d199c7fd291e5f5d95f6ed5c7cf4c2f1df"} Feb 28 11:28:02 crc kubenswrapper[4972]: I0228 11:28:02.624764 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" event={"ID":"a30a4c89-8677-49ac-8709-a63c48017fd0","Type":"ContainerStarted","Data":"ab2253b49a7f77838fec19570769e8f27e4109a778ad1cb3fe123ec30f1170eb"} Feb 28 11:28:03 crc kubenswrapper[4972]: I0228 11:28:03.637101 4972 generic.go:334] "Generic (PLEG): container finished" podID="a30a4c89-8677-49ac-8709-a63c48017fd0" containerID="ab2253b49a7f77838fec19570769e8f27e4109a778ad1cb3fe123ec30f1170eb" exitCode=0 Feb 28 11:28:03 crc kubenswrapper[4972]: I0228 11:28:03.637171 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" event={"ID":"a30a4c89-8677-49ac-8709-a63c48017fd0","Type":"ContainerDied","Data":"ab2253b49a7f77838fec19570769e8f27e4109a778ad1cb3fe123ec30f1170eb"} Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.246709 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.289868 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9h89\" (UniqueName: \"kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89\") pod \"a30a4c89-8677-49ac-8709-a63c48017fd0\" (UID: \"a30a4c89-8677-49ac-8709-a63c48017fd0\") " Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.347929 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89" (OuterVolumeSpecName: "kube-api-access-r9h89") pod "a30a4c89-8677-49ac-8709-a63c48017fd0" (UID: "a30a4c89-8677-49ac-8709-a63c48017fd0"). InnerVolumeSpecName "kube-api-access-r9h89". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.391292 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9h89\" (UniqueName: \"kubernetes.io/projected/a30a4c89-8677-49ac-8709-a63c48017fd0-kube-api-access-r9h89\") on node \"crc\" DevicePath \"\"" Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.668858 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" event={"ID":"a30a4c89-8677-49ac-8709-a63c48017fd0","Type":"ContainerDied","Data":"d5850ee099ec82fd0fb5020b789c42d199c7fd291e5f5d95f6ed5c7cf4c2f1df"} Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.668944 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5850ee099ec82fd0fb5020b789c42d199c7fd291e5f5d95f6ed5c7cf4c2f1df" Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.669103 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537968-l8pzf" Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.743652 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537962-srlzk"] Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.756745 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537962-srlzk"] Feb 28 11:28:05 crc kubenswrapper[4972]: I0228 11:28:05.803352 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9de5d4-56bc-4f9e-a3ef-57925038c881" path="/var/lib/kubelet/pods/5a9de5d4-56bc-4f9e-a3ef-57925038c881/volumes" Feb 28 11:28:10 crc kubenswrapper[4972]: I0228 11:28:10.414222 4972 scope.go:117] "RemoveContainer" containerID="71efbeda45649382ea33e095ccdfc6a4e2d1755714249e8f3f94e481e4bf563a" Feb 28 11:28:28 crc kubenswrapper[4972]: I0228 11:28:28.890966 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:28:28 crc kubenswrapper[4972]: I0228 11:28:28.891848 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:28:58 crc kubenswrapper[4972]: I0228 11:28:58.891336 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:28:58 crc kubenswrapper[4972]: I0228 11:28:58.892094 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:29:28 crc kubenswrapper[4972]: I0228 11:29:28.891268 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:29:28 crc kubenswrapper[4972]: I0228 11:29:28.892378 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:29:28 crc kubenswrapper[4972]: I0228 11:29:28.892482 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:29:28 crc kubenswrapper[4972]: I0228 11:29:28.893941 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:29:28 crc kubenswrapper[4972]: I0228 11:29:28.894726 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" gracePeriod=600 Feb 28 11:29:29 crc kubenswrapper[4972]: E0228 11:29:29.038248 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:29:29 crc kubenswrapper[4972]: I0228 11:29:29.652888 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" exitCode=0 Feb 28 11:29:29 crc kubenswrapper[4972]: I0228 11:29:29.652952 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e"} Feb 28 11:29:29 crc kubenswrapper[4972]: I0228 11:29:29.653012 4972 scope.go:117] "RemoveContainer" containerID="20b52231fcd56296d3d4bcb5a8e98ca69edbaabfc96e7a74302949a765218024" Feb 28 11:29:29 crc kubenswrapper[4972]: I0228 11:29:29.654030 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:29:29 crc kubenswrapper[4972]: E0228 11:29:29.654935 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:29:42 crc kubenswrapper[4972]: I0228 11:29:42.789345 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:29:42 crc kubenswrapper[4972]: E0228 11:29:42.790765 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:29:54 crc kubenswrapper[4972]: I0228 11:29:54.790802 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:29:54 crc kubenswrapper[4972]: E0228 11:29:54.791974 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.199207 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537970-7x4x2"] Feb 28 11:30:00 crc kubenswrapper[4972]: E0228 11:30:00.200938 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30a4c89-8677-49ac-8709-a63c48017fd0" containerName="oc" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.201014 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30a4c89-8677-49ac-8709-a63c48017fd0" containerName="oc" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.201271 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30a4c89-8677-49ac-8709-a63c48017fd0" containerName="oc" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.201949 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.203950 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.204722 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.211310 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.220808 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz"] Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.222190 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.224126 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.232819 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.234928 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537970-7x4x2"] Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.245727 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz"] Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.363189 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzt2l\" (UniqueName: \"kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.363629 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jht6\" (UniqueName: \"kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6\") pod \"auto-csr-approver-29537970-7x4x2\" (UID: \"316ae654-63a3-44b0-b947-300ce18544eb\") " pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.363677 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.363999 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.465698 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzt2l\" (UniqueName: \"kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.465794 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jht6\" (UniqueName: \"kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6\") pod \"auto-csr-approver-29537970-7x4x2\" (UID: \"316ae654-63a3-44b0-b947-300ce18544eb\") " pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.465839 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.465920 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.468330 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.474608 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.489102 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jht6\" (UniqueName: \"kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6\") pod \"auto-csr-approver-29537970-7x4x2\" (UID: \"316ae654-63a3-44b0-b947-300ce18544eb\") " pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.492008 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzt2l\" (UniqueName: \"kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l\") pod \"collect-profiles-29537970-dkxvz\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.578860 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:00 crc kubenswrapper[4972]: I0228 11:30:00.586370 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:01 crc kubenswrapper[4972]: I0228 11:30:01.162002 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537970-7x4x2"] Feb 28 11:30:01 crc kubenswrapper[4972]: I0228 11:30:01.252559 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz"] Feb 28 11:30:01 crc kubenswrapper[4972]: W0228 11:30:01.258876 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11af5fca_81f3_46d3_9b55_0517b585f93e.slice/crio-74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda WatchSource:0}: Error finding container 74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda: Status 404 returned error can't find the container with id 74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda Feb 28 11:30:02 crc kubenswrapper[4972]: I0228 11:30:02.055408 4972 generic.go:334] "Generic (PLEG): container finished" podID="11af5fca-81f3-46d3-9b55-0517b585f93e" containerID="cc76993e43ca905bcd15a9c1d8b3029021c0ec433d89376494a16ed89be71c80" exitCode=0 Feb 28 11:30:02 crc kubenswrapper[4972]: I0228 11:30:02.055541 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" event={"ID":"11af5fca-81f3-46d3-9b55-0517b585f93e","Type":"ContainerDied","Data":"cc76993e43ca905bcd15a9c1d8b3029021c0ec433d89376494a16ed89be71c80"} Feb 28 11:30:02 crc kubenswrapper[4972]: I0228 11:30:02.055880 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" event={"ID":"11af5fca-81f3-46d3-9b55-0517b585f93e","Type":"ContainerStarted","Data":"74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda"} Feb 28 11:30:02 crc kubenswrapper[4972]: I0228 11:30:02.057720 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" event={"ID":"316ae654-63a3-44b0-b947-300ce18544eb","Type":"ContainerStarted","Data":"b440edc3c019d27d87fb77ed8248b364e011bc02124176f0306815cb879da52f"} Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.566384 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.733127 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzt2l\" (UniqueName: \"kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l\") pod \"11af5fca-81f3-46d3-9b55-0517b585f93e\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.733284 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume\") pod \"11af5fca-81f3-46d3-9b55-0517b585f93e\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.733396 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume\") pod \"11af5fca-81f3-46d3-9b55-0517b585f93e\" (UID: \"11af5fca-81f3-46d3-9b55-0517b585f93e\") " Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.734033 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume" (OuterVolumeSpecName: "config-volume") pod "11af5fca-81f3-46d3-9b55-0517b585f93e" (UID: "11af5fca-81f3-46d3-9b55-0517b585f93e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.740194 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l" (OuterVolumeSpecName: "kube-api-access-xzt2l") pod "11af5fca-81f3-46d3-9b55-0517b585f93e" (UID: "11af5fca-81f3-46d3-9b55-0517b585f93e"). InnerVolumeSpecName "kube-api-access-xzt2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.740681 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "11af5fca-81f3-46d3-9b55-0517b585f93e" (UID: "11af5fca-81f3-46d3-9b55-0517b585f93e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.837793 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzt2l\" (UniqueName: \"kubernetes.io/projected/11af5fca-81f3-46d3-9b55-0517b585f93e-kube-api-access-xzt2l\") on node \"crc\" DevicePath \"\"" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.837831 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/11af5fca-81f3-46d3-9b55-0517b585f93e-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:30:03 crc kubenswrapper[4972]: I0228 11:30:03.837845 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11af5fca-81f3-46d3-9b55-0517b585f93e-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.085015 4972 generic.go:334] "Generic (PLEG): container finished" podID="316ae654-63a3-44b0-b947-300ce18544eb" containerID="8f6595d4f6d13a468c6831c50d4c53063d3d9f3bfa278afbfb27e7221c4a7764" exitCode=0 Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.085117 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" event={"ID":"316ae654-63a3-44b0-b947-300ce18544eb","Type":"ContainerDied","Data":"8f6595d4f6d13a468c6831c50d4c53063d3d9f3bfa278afbfb27e7221c4a7764"} Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.089900 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" event={"ID":"11af5fca-81f3-46d3-9b55-0517b585f93e","Type":"ContainerDied","Data":"74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda"} Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.089966 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74a8533116fc150f3a4a56fa64a5a56127823840108d4ae74bbe849065554bda" Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.090047 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537970-dkxvz" Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.667359 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6"] Feb 28 11:30:04 crc kubenswrapper[4972]: I0228 11:30:04.682024 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537925-tgtt6"] Feb 28 11:30:05 crc kubenswrapper[4972]: I0228 11:30:05.577274 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:05 crc kubenswrapper[4972]: I0228 11:30:05.680127 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jht6\" (UniqueName: \"kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6\") pod \"316ae654-63a3-44b0-b947-300ce18544eb\" (UID: \"316ae654-63a3-44b0-b947-300ce18544eb\") " Feb 28 11:30:05 crc kubenswrapper[4972]: I0228 11:30:05.686548 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6" (OuterVolumeSpecName: "kube-api-access-2jht6") pod "316ae654-63a3-44b0-b947-300ce18544eb" (UID: "316ae654-63a3-44b0-b947-300ce18544eb"). InnerVolumeSpecName "kube-api-access-2jht6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:30:05 crc kubenswrapper[4972]: I0228 11:30:05.783190 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jht6\" (UniqueName: \"kubernetes.io/projected/316ae654-63a3-44b0-b947-300ce18544eb-kube-api-access-2jht6\") on node \"crc\" DevicePath \"\"" Feb 28 11:30:05 crc kubenswrapper[4972]: I0228 11:30:05.806558 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8dd23ca-32b0-46f1-a8c7-bb08c70727f6" path="/var/lib/kubelet/pods/c8dd23ca-32b0-46f1-a8c7-bb08c70727f6/volumes" Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.118164 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" event={"ID":"316ae654-63a3-44b0-b947-300ce18544eb","Type":"ContainerDied","Data":"b440edc3c019d27d87fb77ed8248b364e011bc02124176f0306815cb879da52f"} Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.118651 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b440edc3c019d27d87fb77ed8248b364e011bc02124176f0306815cb879da52f" Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.118270 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537970-7x4x2" Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.654177 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537964-2cbzx"] Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.663722 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537964-2cbzx"] Feb 28 11:30:06 crc kubenswrapper[4972]: I0228 11:30:06.790373 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:30:06 crc kubenswrapper[4972]: E0228 11:30:06.790876 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:30:07 crc kubenswrapper[4972]: I0228 11:30:07.803661 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe5ae58-c5f0-4540-aa4a-191bbde11f2a" path="/var/lib/kubelet/pods/7fe5ae58-c5f0-4540-aa4a-191bbde11f2a/volumes" Feb 28 11:30:10 crc kubenswrapper[4972]: I0228 11:30:10.587243 4972 scope.go:117] "RemoveContainer" containerID="471517a922347f1134b014ba802be3250743788a53d43efe6bbc797d79162ee7" Feb 28 11:30:10 crc kubenswrapper[4972]: I0228 11:30:10.626746 4972 scope.go:117] "RemoveContainer" containerID="2f0f99b14f7575771ab0150c659d1ef9b508ada9404924c7544bbc509a64cf08" Feb 28 11:30:17 crc kubenswrapper[4972]: I0228 11:30:17.790483 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:30:17 crc kubenswrapper[4972]: E0228 11:30:17.792117 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:30:29 crc kubenswrapper[4972]: I0228 11:30:29.789669 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:30:29 crc kubenswrapper[4972]: E0228 11:30:29.791037 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:30:41 crc kubenswrapper[4972]: I0228 11:30:41.789263 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:30:41 crc kubenswrapper[4972]: E0228 11:30:41.790182 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:30:52 crc kubenswrapper[4972]: I0228 11:30:52.789960 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:30:52 crc kubenswrapper[4972]: E0228 11:30:52.792582 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:31:06 crc kubenswrapper[4972]: I0228 11:31:06.790417 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:31:06 crc kubenswrapper[4972]: E0228 11:31:06.791601 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:31:21 crc kubenswrapper[4972]: I0228 11:31:21.790979 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:31:21 crc kubenswrapper[4972]: E0228 11:31:21.792290 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:31:33 crc kubenswrapper[4972]: I0228 11:31:33.795418 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:31:33 crc kubenswrapper[4972]: E0228 11:31:33.796611 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:31:44 crc kubenswrapper[4972]: I0228 11:31:44.790588 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:31:44 crc kubenswrapper[4972]: E0228 11:31:44.791929 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:31:58 crc kubenswrapper[4972]: I0228 11:31:58.789644 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:31:58 crc kubenswrapper[4972]: E0228 11:31:58.790797 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.161537 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537972-t9k8k"] Feb 28 11:32:00 crc kubenswrapper[4972]: E0228 11:32:00.162514 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11af5fca-81f3-46d3-9b55-0517b585f93e" containerName="collect-profiles" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.162535 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="11af5fca-81f3-46d3-9b55-0517b585f93e" containerName="collect-profiles" Feb 28 11:32:00 crc kubenswrapper[4972]: E0228 11:32:00.162571 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316ae654-63a3-44b0-b947-300ce18544eb" containerName="oc" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.162583 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="316ae654-63a3-44b0-b947-300ce18544eb" containerName="oc" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.162906 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="316ae654-63a3-44b0-b947-300ce18544eb" containerName="oc" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.162947 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="11af5fca-81f3-46d3-9b55-0517b585f93e" containerName="collect-profiles" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.164161 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.167084 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.167443 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.170032 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.182586 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537972-t9k8k"] Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.276563 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mpjv\" (UniqueName: \"kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv\") pod \"auto-csr-approver-29537972-t9k8k\" (UID: \"e3c4beba-180c-45c5-9012-f3a483a9844b\") " pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.379249 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mpjv\" (UniqueName: \"kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv\") pod \"auto-csr-approver-29537972-t9k8k\" (UID: \"e3c4beba-180c-45c5-9012-f3a483a9844b\") " pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.436427 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mpjv\" (UniqueName: \"kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv\") pod \"auto-csr-approver-29537972-t9k8k\" (UID: \"e3c4beba-180c-45c5-9012-f3a483a9844b\") " pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:00 crc kubenswrapper[4972]: I0228 11:32:00.498197 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:01 crc kubenswrapper[4972]: I0228 11:32:01.092705 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537972-t9k8k"] Feb 28 11:32:01 crc kubenswrapper[4972]: I0228 11:32:01.100578 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:32:01 crc kubenswrapper[4972]: I0228 11:32:01.415032 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" event={"ID":"e3c4beba-180c-45c5-9012-f3a483a9844b","Type":"ContainerStarted","Data":"c2325bfe787ffeaa1e094d82f67f3c6349ad17886f390dc0cbf7368647b279f4"} Feb 28 11:32:02 crc kubenswrapper[4972]: I0228 11:32:02.431787 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" event={"ID":"e3c4beba-180c-45c5-9012-f3a483a9844b","Type":"ContainerStarted","Data":"56d84b1356b7569072a8aa9c55752e41664847d1f01e7fd637010b8660455d1e"} Feb 28 11:32:02 crc kubenswrapper[4972]: I0228 11:32:02.451889 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" podStartSLOduration=1.615812872 podStartE2EDuration="2.451869395s" podCreationTimestamp="2026-02-28 11:32:00 +0000 UTC" firstStartedPulling="2026-02-28 11:32:01.100195032 +0000 UTC m=+3378.012202810" lastFinishedPulling="2026-02-28 11:32:01.936251595 +0000 UTC m=+3378.848259333" observedRunningTime="2026-02-28 11:32:02.444912985 +0000 UTC m=+3379.356920733" watchObservedRunningTime="2026-02-28 11:32:02.451869395 +0000 UTC m=+3379.363877153" Feb 28 11:32:03 crc kubenswrapper[4972]: I0228 11:32:03.442897 4972 generic.go:334] "Generic (PLEG): container finished" podID="e3c4beba-180c-45c5-9012-f3a483a9844b" containerID="56d84b1356b7569072a8aa9c55752e41664847d1f01e7fd637010b8660455d1e" exitCode=0 Feb 28 11:32:03 crc kubenswrapper[4972]: I0228 11:32:03.442983 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" event={"ID":"e3c4beba-180c-45c5-9012-f3a483a9844b","Type":"ContainerDied","Data":"56d84b1356b7569072a8aa9c55752e41664847d1f01e7fd637010b8660455d1e"} Feb 28 11:32:04 crc kubenswrapper[4972]: I0228 11:32:04.955144 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.094812 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mpjv\" (UniqueName: \"kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv\") pod \"e3c4beba-180c-45c5-9012-f3a483a9844b\" (UID: \"e3c4beba-180c-45c5-9012-f3a483a9844b\") " Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.104341 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv" (OuterVolumeSpecName: "kube-api-access-9mpjv") pod "e3c4beba-180c-45c5-9012-f3a483a9844b" (UID: "e3c4beba-180c-45c5-9012-f3a483a9844b"). InnerVolumeSpecName "kube-api-access-9mpjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.197833 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mpjv\" (UniqueName: \"kubernetes.io/projected/e3c4beba-180c-45c5-9012-f3a483a9844b-kube-api-access-9mpjv\") on node \"crc\" DevicePath \"\"" Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.507629 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" event={"ID":"e3c4beba-180c-45c5-9012-f3a483a9844b","Type":"ContainerDied","Data":"c2325bfe787ffeaa1e094d82f67f3c6349ad17886f390dc0cbf7368647b279f4"} Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.507673 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2325bfe787ffeaa1e094d82f67f3c6349ad17886f390dc0cbf7368647b279f4" Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.507728 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537972-t9k8k" Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.558349 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537966-2bwd2"] Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.566979 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537966-2bwd2"] Feb 28 11:32:05 crc kubenswrapper[4972]: I0228 11:32:05.807169 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="357dbfdb-c40c-49b6-a1d2-ffdca782fe7c" path="/var/lib/kubelet/pods/357dbfdb-c40c-49b6-a1d2-ffdca782fe7c/volumes" Feb 28 11:32:10 crc kubenswrapper[4972]: I0228 11:32:10.787526 4972 scope.go:117] "RemoveContainer" containerID="a06463251c2f039b055f84d90a36ca15275586d0c8a13ab28c58358cd09932d4" Feb 28 11:32:13 crc kubenswrapper[4972]: I0228 11:32:13.840748 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:32:13 crc kubenswrapper[4972]: E0228 11:32:13.858935 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:32:29 crc kubenswrapper[4972]: I0228 11:32:29.789987 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:32:29 crc kubenswrapper[4972]: E0228 11:32:29.790810 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:32:40 crc kubenswrapper[4972]: I0228 11:32:40.790414 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:32:40 crc kubenswrapper[4972]: E0228 11:32:40.792135 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:32:53 crc kubenswrapper[4972]: I0228 11:32:53.797517 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:32:53 crc kubenswrapper[4972]: E0228 11:32:53.798671 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:33:08 crc kubenswrapper[4972]: I0228 11:33:08.789950 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:33:08 crc kubenswrapper[4972]: E0228 11:33:08.791123 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:33:21 crc kubenswrapper[4972]: I0228 11:33:21.790505 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:33:21 crc kubenswrapper[4972]: E0228 11:33:21.792036 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:33:30 crc kubenswrapper[4972]: I0228 11:33:30.544106 4972 generic.go:334] "Generic (PLEG): container finished" podID="50684e40-2fa2-4cb3-a176-374ab716698c" containerID="5aae7280aa24c06af05541d09a1032228451f1ae547d9db972bc67bce2aa9374" exitCode=0 Feb 28 11:33:30 crc kubenswrapper[4972]: I0228 11:33:30.544240 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"50684e40-2fa2-4cb3-a176-374ab716698c","Type":"ContainerDied","Data":"5aae7280aa24c06af05541d09a1032228451f1ae547d9db972bc67bce2aa9374"} Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.034084 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154162 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154232 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154299 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154386 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwwg6\" (UniqueName: \"kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154523 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154568 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154680 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154761 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.154803 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret\") pod \"50684e40-2fa2-4cb3-a176-374ab716698c\" (UID: \"50684e40-2fa2-4cb3-a176-374ab716698c\") " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.155939 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.156275 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data" (OuterVolumeSpecName: "config-data") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.162230 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6" (OuterVolumeSpecName: "kube-api-access-jwwg6") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "kube-api-access-jwwg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.164448 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.177365 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.194279 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.212534 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.215484 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.217872 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "50684e40-2fa2-4cb3-a176-374ab716698c" (UID: "50684e40-2fa2-4cb3-a176-374ab716698c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265730 4972 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-config-data\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265772 4972 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265795 4972 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265812 4972 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265831 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265851 4972 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/50684e40-2fa2-4cb3-a176-374ab716698c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.265867 4972 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50684e40-2fa2-4cb3-a176-374ab716698c-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.269552 4972 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.273637 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwwg6\" (UniqueName: \"kubernetes.io/projected/50684e40-2fa2-4cb3-a176-374ab716698c-kube-api-access-jwwg6\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.300874 4972 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.376167 4972 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.569958 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"50684e40-2fa2-4cb3-a176-374ab716698c","Type":"ContainerDied","Data":"94eddbd462fbbd91895e6276cefcf32288bfeb0bf8d504e85b43b329e8036d2e"} Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.569995 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94eddbd462fbbd91895e6276cefcf32288bfeb0bf8d504e85b43b329e8036d2e" Feb 28 11:33:32 crc kubenswrapper[4972]: I0228 11:33:32.570085 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 28 11:33:34 crc kubenswrapper[4972]: I0228 11:33:34.789686 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:33:34 crc kubenswrapper[4972]: E0228 11:33:34.790486 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.707020 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 28 11:33:43 crc kubenswrapper[4972]: E0228 11:33:43.708543 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50684e40-2fa2-4cb3-a176-374ab716698c" containerName="tempest-tests-tempest-tests-runner" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.708564 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="50684e40-2fa2-4cb3-a176-374ab716698c" containerName="tempest-tests-tempest-tests-runner" Feb 28 11:33:43 crc kubenswrapper[4972]: E0228 11:33:43.708609 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c4beba-180c-45c5-9012-f3a483a9844b" containerName="oc" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.708620 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c4beba-180c-45c5-9012-f3a483a9844b" containerName="oc" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.709185 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c4beba-180c-45c5-9012-f3a483a9844b" containerName="oc" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.709256 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="50684e40-2fa2-4cb3-a176-374ab716698c" containerName="tempest-tests-tempest-tests-runner" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.710563 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.712720 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vgjz8" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.732263 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.849963 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.850073 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x52f\" (UniqueName: \"kubernetes.io/projected/fb460ac4-a796-4700-ab43-565ca0308a20-kube-api-access-6x52f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.953113 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.953353 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x52f\" (UniqueName: \"kubernetes.io/projected/fb460ac4-a796-4700-ab43-565ca0308a20-kube-api-access-6x52f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.954501 4972 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:43 crc kubenswrapper[4972]: I0228 11:33:43.987525 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x52f\" (UniqueName: \"kubernetes.io/projected/fb460ac4-a796-4700-ab43-565ca0308a20-kube-api-access-6x52f\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:44 crc kubenswrapper[4972]: I0228 11:33:44.007099 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fb460ac4-a796-4700-ab43-565ca0308a20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:44 crc kubenswrapper[4972]: I0228 11:33:44.040591 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vgjz8" Feb 28 11:33:44 crc kubenswrapper[4972]: I0228 11:33:44.049518 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 28 11:33:44 crc kubenswrapper[4972]: I0228 11:33:44.389678 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 28 11:33:44 crc kubenswrapper[4972]: I0228 11:33:44.732808 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fb460ac4-a796-4700-ab43-565ca0308a20","Type":"ContainerStarted","Data":"434b41c2ca47e1b0956f453fbbecd7a16176e2fa6860191af535ce812aaa784e"} Feb 28 11:33:45 crc kubenswrapper[4972]: I0228 11:33:45.745028 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fb460ac4-a796-4700-ab43-565ca0308a20","Type":"ContainerStarted","Data":"c7280488e8dbb61fa463d3111a2943dd2d2bd320cdb4dddf303d7f7a00582517"} Feb 28 11:33:45 crc kubenswrapper[4972]: I0228 11:33:45.773288 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.8188721270000001 podStartE2EDuration="2.773268604s" podCreationTimestamp="2026-02-28 11:33:43 +0000 UTC" firstStartedPulling="2026-02-28 11:33:44.377916481 +0000 UTC m=+3481.289924269" lastFinishedPulling="2026-02-28 11:33:45.332312978 +0000 UTC m=+3482.244320746" observedRunningTime="2026-02-28 11:33:45.764799723 +0000 UTC m=+3482.676807501" watchObservedRunningTime="2026-02-28 11:33:45.773268604 +0000 UTC m=+3482.685276352" Feb 28 11:33:47 crc kubenswrapper[4972]: I0228 11:33:47.789751 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:33:47 crc kubenswrapper[4972]: E0228 11:33:47.790269 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.202351 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537974-pr6fq"] Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.207151 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.218542 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537974-pr6fq"] Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.222547 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.222886 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.223219 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.239073 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm2rc\" (UniqueName: \"kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc\") pod \"auto-csr-approver-29537974-pr6fq\" (UID: \"31243ab6-69bc-46b0-83b9-da39f9f11074\") " pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.342060 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm2rc\" (UniqueName: \"kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc\") pod \"auto-csr-approver-29537974-pr6fq\" (UID: \"31243ab6-69bc-46b0-83b9-da39f9f11074\") " pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.370318 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm2rc\" (UniqueName: \"kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc\") pod \"auto-csr-approver-29537974-pr6fq\" (UID: \"31243ab6-69bc-46b0-83b9-da39f9f11074\") " pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.543085 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:00 crc kubenswrapper[4972]: I0228 11:34:00.789418 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:34:00 crc kubenswrapper[4972]: E0228 11:34:00.789838 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:34:01 crc kubenswrapper[4972]: I0228 11:34:01.074073 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537974-pr6fq"] Feb 28 11:34:01 crc kubenswrapper[4972]: I0228 11:34:01.967158 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" event={"ID":"31243ab6-69bc-46b0-83b9-da39f9f11074","Type":"ContainerStarted","Data":"67925671717a24f42c33b4a29e1a9f9f17740cb7ac3dca146bff2c202ff4ac11"} Feb 28 11:34:02 crc kubenswrapper[4972]: I0228 11:34:02.982601 4972 generic.go:334] "Generic (PLEG): container finished" podID="31243ab6-69bc-46b0-83b9-da39f9f11074" containerID="9204d1f38c271e8afe07636ec07bfb9355ae86603b6561193a9ef611185b0245" exitCode=0 Feb 28 11:34:02 crc kubenswrapper[4972]: I0228 11:34:02.982683 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" event={"ID":"31243ab6-69bc-46b0-83b9-da39f9f11074","Type":"ContainerDied","Data":"9204d1f38c271e8afe07636ec07bfb9355ae86603b6561193a9ef611185b0245"} Feb 28 11:34:04 crc kubenswrapper[4972]: I0228 11:34:04.468447 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:04 crc kubenswrapper[4972]: I0228 11:34:04.540591 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm2rc\" (UniqueName: \"kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc\") pod \"31243ab6-69bc-46b0-83b9-da39f9f11074\" (UID: \"31243ab6-69bc-46b0-83b9-da39f9f11074\") " Feb 28 11:34:04 crc kubenswrapper[4972]: I0228 11:34:04.548911 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc" (OuterVolumeSpecName: "kube-api-access-sm2rc") pod "31243ab6-69bc-46b0-83b9-da39f9f11074" (UID: "31243ab6-69bc-46b0-83b9-da39f9f11074"). InnerVolumeSpecName "kube-api-access-sm2rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:34:04 crc kubenswrapper[4972]: I0228 11:34:04.653631 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm2rc\" (UniqueName: \"kubernetes.io/projected/31243ab6-69bc-46b0-83b9-da39f9f11074-kube-api-access-sm2rc\") on node \"crc\" DevicePath \"\"" Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.012358 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" event={"ID":"31243ab6-69bc-46b0-83b9-da39f9f11074","Type":"ContainerDied","Data":"67925671717a24f42c33b4a29e1a9f9f17740cb7ac3dca146bff2c202ff4ac11"} Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.012828 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67925671717a24f42c33b4a29e1a9f9f17740cb7ac3dca146bff2c202ff4ac11" Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.012646 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537974-pr6fq" Feb 28 11:34:05 crc kubenswrapper[4972]: E0228 11:34:05.236979 4972 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31243ab6_69bc_46b0_83b9_da39f9f11074.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31243ab6_69bc_46b0_83b9_da39f9f11074.slice/crio-67925671717a24f42c33b4a29e1a9f9f17740cb7ac3dca146bff2c202ff4ac11\": RecentStats: unable to find data in memory cache]" Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.549412 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537968-l8pzf"] Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.560621 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537968-l8pzf"] Feb 28 11:34:05 crc kubenswrapper[4972]: I0228 11:34:05.811551 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a30a4c89-8677-49ac-8709-a63c48017fd0" path="/var/lib/kubelet/pods/a30a4c89-8677-49ac-8709-a63c48017fd0/volumes" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.396519 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jmjz5/must-gather-s544x"] Feb 28 11:34:07 crc kubenswrapper[4972]: E0228 11:34:07.397186 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31243ab6-69bc-46b0-83b9-da39f9f11074" containerName="oc" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.397208 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="31243ab6-69bc-46b0-83b9-da39f9f11074" containerName="oc" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.397534 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="31243ab6-69bc-46b0-83b9-da39f9f11074" containerName="oc" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.399150 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.401677 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jmjz5"/"openshift-service-ca.crt" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.401995 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jmjz5"/"default-dockercfg-bdtqv" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.402767 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jmjz5"/"kube-root-ca.crt" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.423709 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.424102 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrzvp\" (UniqueName: \"kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.504503 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jmjz5/must-gather-s544x"] Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.526416 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrzvp\" (UniqueName: \"kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.526614 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.527118 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.550430 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrzvp\" (UniqueName: \"kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp\") pod \"must-gather-s544x\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:07 crc kubenswrapper[4972]: I0228 11:34:07.726377 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:34:08 crc kubenswrapper[4972]: I0228 11:34:08.203027 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jmjz5/must-gather-s544x"] Feb 28 11:34:08 crc kubenswrapper[4972]: W0228 11:34:08.205694 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9b9a51d_5064_4502_a8fc_3a1cc75d591b.slice/crio-1f70d1ff76802cce6b0702778dff204c6033c993eeb49abec15a1a66b6a02214 WatchSource:0}: Error finding container 1f70d1ff76802cce6b0702778dff204c6033c993eeb49abec15a1a66b6a02214: Status 404 returned error can't find the container with id 1f70d1ff76802cce6b0702778dff204c6033c993eeb49abec15a1a66b6a02214 Feb 28 11:34:09 crc kubenswrapper[4972]: I0228 11:34:09.062450 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/must-gather-s544x" event={"ID":"c9b9a51d-5064-4502-a8fc-3a1cc75d591b","Type":"ContainerStarted","Data":"1f70d1ff76802cce6b0702778dff204c6033c993eeb49abec15a1a66b6a02214"} Feb 28 11:34:10 crc kubenswrapper[4972]: I0228 11:34:10.923903 4972 scope.go:117] "RemoveContainer" containerID="ab2253b49a7f77838fec19570769e8f27e4109a778ad1cb3fe123ec30f1170eb" Feb 28 11:34:11 crc kubenswrapper[4972]: I0228 11:34:11.796263 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:34:11 crc kubenswrapper[4972]: E0228 11:34:11.797283 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:34:16 crc kubenswrapper[4972]: I0228 11:34:16.216621 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/must-gather-s544x" event={"ID":"c9b9a51d-5064-4502-a8fc-3a1cc75d591b","Type":"ContainerStarted","Data":"e8e5d4c2262051ebb4fba44c6bb5c34add63d6c9f1f0482324754e65cb9f1abd"} Feb 28 11:34:16 crc kubenswrapper[4972]: I0228 11:34:16.216980 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/must-gather-s544x" event={"ID":"c9b9a51d-5064-4502-a8fc-3a1cc75d591b","Type":"ContainerStarted","Data":"14e22614dc3ff190e17c999c58a137372f7236e13b44de008bd27336ef3f4846"} Feb 28 11:34:16 crc kubenswrapper[4972]: I0228 11:34:16.245659 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jmjz5/must-gather-s544x" podStartSLOduration=2.072304932 podStartE2EDuration="9.245636173s" podCreationTimestamp="2026-02-28 11:34:07 +0000 UTC" firstStartedPulling="2026-02-28 11:34:08.20944747 +0000 UTC m=+3505.121455218" lastFinishedPulling="2026-02-28 11:34:15.382778721 +0000 UTC m=+3512.294786459" observedRunningTime="2026-02-28 11:34:16.244291156 +0000 UTC m=+3513.156298934" watchObservedRunningTime="2026-02-28 11:34:16.245636173 +0000 UTC m=+3513.157643951" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.838802 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-lngr4"] Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.840501 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.876818 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s966\" (UniqueName: \"kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.879595 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.982200 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.982405 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:19 crc kubenswrapper[4972]: I0228 11:34:19.982432 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s966\" (UniqueName: \"kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:20 crc kubenswrapper[4972]: I0228 11:34:20.011638 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s966\" (UniqueName: \"kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966\") pod \"crc-debug-lngr4\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:20 crc kubenswrapper[4972]: I0228 11:34:20.163909 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:34:20 crc kubenswrapper[4972]: I0228 11:34:20.269057 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" event={"ID":"5ce7f480-9ce3-4ade-98e5-133181b37425","Type":"ContainerStarted","Data":"717a4daa0d54e407391b9f4979429f47ace0da79ab64b304b96887f62097ec44"} Feb 28 11:34:22 crc kubenswrapper[4972]: I0228 11:34:22.790125 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:34:22 crc kubenswrapper[4972]: E0228 11:34:22.790859 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:34:33 crc kubenswrapper[4972]: I0228 11:34:33.404187 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" event={"ID":"5ce7f480-9ce3-4ade-98e5-133181b37425","Type":"ContainerStarted","Data":"77a679720e8485b9d9503de24014051b6f1912c02b5a8740ae8962f1ae60aa6b"} Feb 28 11:34:33 crc kubenswrapper[4972]: I0228 11:34:33.422570 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" podStartSLOduration=1.594783285 podStartE2EDuration="14.422551916s" podCreationTimestamp="2026-02-28 11:34:19 +0000 UTC" firstStartedPulling="2026-02-28 11:34:20.188632927 +0000 UTC m=+3517.100640665" lastFinishedPulling="2026-02-28 11:34:33.016401558 +0000 UTC m=+3529.928409296" observedRunningTime="2026-02-28 11:34:33.417880469 +0000 UTC m=+3530.329888207" watchObservedRunningTime="2026-02-28 11:34:33.422551916 +0000 UTC m=+3530.334559654" Feb 28 11:34:37 crc kubenswrapper[4972]: I0228 11:34:37.790371 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:34:41 crc kubenswrapper[4972]: I0228 11:34:41.488498 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7"} Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.672343 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.674632 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.687804 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.865054 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.865120 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbpvr\" (UniqueName: \"kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.865153 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.967273 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbpvr\" (UniqueName: \"kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.967351 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.967586 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.968725 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.969502 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:55 crc kubenswrapper[4972]: I0228 11:34:55.989501 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbpvr\" (UniqueName: \"kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr\") pod \"redhat-marketplace-8sbkc\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:56 crc kubenswrapper[4972]: I0228 11:34:56.045995 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:34:56 crc kubenswrapper[4972]: I0228 11:34:56.576868 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:34:56 crc kubenswrapper[4972]: I0228 11:34:56.642183 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerStarted","Data":"364c86787ef6fa5c7062a0cf8dd00804f3a67dda39599eef0fbab3a43a8b6cc7"} Feb 28 11:34:57 crc kubenswrapper[4972]: I0228 11:34:57.656969 4972 generic.go:334] "Generic (PLEG): container finished" podID="09f3fded-90f5-4410-a9d4-700118e399ea" containerID="972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70" exitCode=0 Feb 28 11:34:57 crc kubenswrapper[4972]: I0228 11:34:57.657270 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerDied","Data":"972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70"} Feb 28 11:34:58 crc kubenswrapper[4972]: I0228 11:34:58.670652 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerStarted","Data":"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f"} Feb 28 11:34:59 crc kubenswrapper[4972]: I0228 11:34:59.689937 4972 generic.go:334] "Generic (PLEG): container finished" podID="09f3fded-90f5-4410-a9d4-700118e399ea" containerID="a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f" exitCode=0 Feb 28 11:34:59 crc kubenswrapper[4972]: I0228 11:34:59.690035 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerDied","Data":"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f"} Feb 28 11:35:00 crc kubenswrapper[4972]: I0228 11:35:00.709839 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerStarted","Data":"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8"} Feb 28 11:35:00 crc kubenswrapper[4972]: I0228 11:35:00.753887 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8sbkc" podStartSLOduration=3.3042877219999998 podStartE2EDuration="5.753867082s" podCreationTimestamp="2026-02-28 11:34:55 +0000 UTC" firstStartedPulling="2026-02-28 11:34:57.659121901 +0000 UTC m=+3554.571129679" lastFinishedPulling="2026-02-28 11:35:00.108701301 +0000 UTC m=+3557.020709039" observedRunningTime="2026-02-28 11:35:00.747556429 +0000 UTC m=+3557.659564167" watchObservedRunningTime="2026-02-28 11:35:00.753867082 +0000 UTC m=+3557.665874830" Feb 28 11:35:06 crc kubenswrapper[4972]: I0228 11:35:06.046294 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:06 crc kubenswrapper[4972]: I0228 11:35:06.048202 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:07 crc kubenswrapper[4972]: I0228 11:35:07.111096 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-8sbkc" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="registry-server" probeResult="failure" output=< Feb 28 11:35:07 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:35:07 crc kubenswrapper[4972]: > Feb 28 11:35:12 crc kubenswrapper[4972]: I0228 11:35:12.859440 4972 generic.go:334] "Generic (PLEG): container finished" podID="5ce7f480-9ce3-4ade-98e5-133181b37425" containerID="77a679720e8485b9d9503de24014051b6f1912c02b5a8740ae8962f1ae60aa6b" exitCode=0 Feb 28 11:35:12 crc kubenswrapper[4972]: I0228 11:35:12.859661 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" event={"ID":"5ce7f480-9ce3-4ade-98e5-133181b37425","Type":"ContainerDied","Data":"77a679720e8485b9d9503de24014051b6f1912c02b5a8740ae8962f1ae60aa6b"} Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.012935 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.058886 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s966\" (UniqueName: \"kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966\") pod \"5ce7f480-9ce3-4ade-98e5-133181b37425\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.059202 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host\") pod \"5ce7f480-9ce3-4ade-98e5-133181b37425\" (UID: \"5ce7f480-9ce3-4ade-98e5-133181b37425\") " Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.060064 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host" (OuterVolumeSpecName: "host") pod "5ce7f480-9ce3-4ade-98e5-133181b37425" (UID: "5ce7f480-9ce3-4ade-98e5-133181b37425"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.070075 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-lngr4"] Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.072889 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966" (OuterVolumeSpecName: "kube-api-access-5s966") pod "5ce7f480-9ce3-4ade-98e5-133181b37425" (UID: "5ce7f480-9ce3-4ade-98e5-133181b37425"). InnerVolumeSpecName "kube-api-access-5s966". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.087360 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-lngr4"] Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.162160 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s966\" (UniqueName: \"kubernetes.io/projected/5ce7f480-9ce3-4ade-98e5-133181b37425-kube-api-access-5s966\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.162210 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5ce7f480-9ce3-4ade-98e5-133181b37425-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.890529 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="717a4daa0d54e407391b9f4979429f47ace0da79ab64b304b96887f62097ec44" Feb 28 11:35:14 crc kubenswrapper[4972]: I0228 11:35:14.890631 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-lngr4" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.303963 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-q9pxn"] Feb 28 11:35:15 crc kubenswrapper[4972]: E0228 11:35:15.304781 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce7f480-9ce3-4ade-98e5-133181b37425" containerName="container-00" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.304815 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce7f480-9ce3-4ade-98e5-133181b37425" containerName="container-00" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.305373 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce7f480-9ce3-4ade-98e5-133181b37425" containerName="container-00" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.306634 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.393230 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.393523 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh5qr\" (UniqueName: \"kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.494945 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh5qr\" (UniqueName: \"kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.495089 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.495232 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.521367 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh5qr\" (UniqueName: \"kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr\") pod \"crc-debug-q9pxn\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.636858 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:15 crc kubenswrapper[4972]: W0228 11:35:15.681301 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5eb633d_57f6_4a83_b69e_740ee8ac99d6.slice/crio-16d3ccdccd1cc5d118da2e8a511d056acf3df2f989c72571e6470d6127f03846 WatchSource:0}: Error finding container 16d3ccdccd1cc5d118da2e8a511d056acf3df2f989c72571e6470d6127f03846: Status 404 returned error can't find the container with id 16d3ccdccd1cc5d118da2e8a511d056acf3df2f989c72571e6470d6127f03846 Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.814002 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce7f480-9ce3-4ade-98e5-133181b37425" path="/var/lib/kubelet/pods/5ce7f480-9ce3-4ade-98e5-133181b37425/volumes" Feb 28 11:35:15 crc kubenswrapper[4972]: I0228 11:35:15.909049 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" event={"ID":"a5eb633d-57f6-4a83-b69e-740ee8ac99d6","Type":"ContainerStarted","Data":"16d3ccdccd1cc5d118da2e8a511d056acf3df2f989c72571e6470d6127f03846"} Feb 28 11:35:16 crc kubenswrapper[4972]: I0228 11:35:16.119534 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:16 crc kubenswrapper[4972]: I0228 11:35:16.210103 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:16 crc kubenswrapper[4972]: I0228 11:35:16.377179 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:35:16 crc kubenswrapper[4972]: I0228 11:35:16.922891 4972 generic.go:334] "Generic (PLEG): container finished" podID="a5eb633d-57f6-4a83-b69e-740ee8ac99d6" containerID="06a6bb3317bf169051e0f91b072468fc774ce3074d6e07c13d4626e7937aaa9a" exitCode=0 Feb 28 11:35:16 crc kubenswrapper[4972]: I0228 11:35:16.922974 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" event={"ID":"a5eb633d-57f6-4a83-b69e-740ee8ac99d6","Type":"ContainerDied","Data":"06a6bb3317bf169051e0f91b072468fc774ce3074d6e07c13d4626e7937aaa9a"} Feb 28 11:35:17 crc kubenswrapper[4972]: I0228 11:35:17.481055 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-q9pxn"] Feb 28 11:35:17 crc kubenswrapper[4972]: I0228 11:35:17.488423 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-q9pxn"] Feb 28 11:35:17 crc kubenswrapper[4972]: I0228 11:35:17.932149 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8sbkc" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="registry-server" containerID="cri-o://5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8" gracePeriod=2 Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.129553 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.277216 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh5qr\" (UniqueName: \"kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr\") pod \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.277279 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host\") pod \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\" (UID: \"a5eb633d-57f6-4a83-b69e-740ee8ac99d6\") " Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.277627 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host" (OuterVolumeSpecName: "host") pod "a5eb633d-57f6-4a83-b69e-740ee8ac99d6" (UID: "a5eb633d-57f6-4a83-b69e-740ee8ac99d6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.293701 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr" (OuterVolumeSpecName: "kube-api-access-mh5qr") pod "a5eb633d-57f6-4a83-b69e-740ee8ac99d6" (UID: "a5eb633d-57f6-4a83-b69e-740ee8ac99d6"). InnerVolumeSpecName "kube-api-access-mh5qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.379744 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh5qr\" (UniqueName: \"kubernetes.io/projected/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-kube-api-access-mh5qr\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.379775 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a5eb633d-57f6-4a83-b69e-740ee8ac99d6-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.390668 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.481425 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbpvr\" (UniqueName: \"kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr\") pod \"09f3fded-90f5-4410-a9d4-700118e399ea\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.482072 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content\") pod \"09f3fded-90f5-4410-a9d4-700118e399ea\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.482145 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities\") pod \"09f3fded-90f5-4410-a9d4-700118e399ea\" (UID: \"09f3fded-90f5-4410-a9d4-700118e399ea\") " Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.482723 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities" (OuterVolumeSpecName: "utilities") pod "09f3fded-90f5-4410-a9d4-700118e399ea" (UID: "09f3fded-90f5-4410-a9d4-700118e399ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.484821 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr" (OuterVolumeSpecName: "kube-api-access-jbpvr") pod "09f3fded-90f5-4410-a9d4-700118e399ea" (UID: "09f3fded-90f5-4410-a9d4-700118e399ea"). InnerVolumeSpecName "kube-api-access-jbpvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.511806 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09f3fded-90f5-4410-a9d4-700118e399ea" (UID: "09f3fded-90f5-4410-a9d4-700118e399ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.584836 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.584892 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09f3fded-90f5-4410-a9d4-700118e399ea-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.584903 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbpvr\" (UniqueName: \"kubernetes.io/projected/09f3fded-90f5-4410-a9d4-700118e399ea-kube-api-access-jbpvr\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.672212 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-gzg78"] Feb 28 11:35:18 crc kubenswrapper[4972]: E0228 11:35:18.672896 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5eb633d-57f6-4a83-b69e-740ee8ac99d6" containerName="container-00" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.672922 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5eb633d-57f6-4a83-b69e-740ee8ac99d6" containerName="container-00" Feb 28 11:35:18 crc kubenswrapper[4972]: E0228 11:35:18.672935 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="registry-server" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.672943 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="registry-server" Feb 28 11:35:18 crc kubenswrapper[4972]: E0228 11:35:18.672962 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="extract-content" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.672969 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="extract-content" Feb 28 11:35:18 crc kubenswrapper[4972]: E0228 11:35:18.672995 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="extract-utilities" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.673005 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="extract-utilities" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.673217 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" containerName="registry-server" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.673250 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5eb633d-57f6-4a83-b69e-740ee8ac99d6" containerName="container-00" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.674041 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.788904 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.789535 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svzt2\" (UniqueName: \"kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.891592 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.891655 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svzt2\" (UniqueName: \"kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.891761 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.911287 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svzt2\" (UniqueName: \"kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2\") pod \"crc-debug-gzg78\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.944902 4972 scope.go:117] "RemoveContainer" containerID="06a6bb3317bf169051e0f91b072468fc774ce3074d6e07c13d4626e7937aaa9a" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.944945 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-q9pxn" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.948231 4972 generic.go:334] "Generic (PLEG): container finished" podID="09f3fded-90f5-4410-a9d4-700118e399ea" containerID="5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8" exitCode=0 Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.948262 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerDied","Data":"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8"} Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.948284 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8sbkc" event={"ID":"09f3fded-90f5-4410-a9d4-700118e399ea","Type":"ContainerDied","Data":"364c86787ef6fa5c7062a0cf8dd00804f3a67dda39599eef0fbab3a43a8b6cc7"} Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.948324 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8sbkc" Feb 28 11:35:18 crc kubenswrapper[4972]: I0228 11:35:18.994299 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.028562 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.030432 4972 scope.go:117] "RemoveContainer" containerID="5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.073376 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8sbkc"] Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.113592 4972 scope.go:117] "RemoveContainer" containerID="a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.175321 4972 scope.go:117] "RemoveContainer" containerID="972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.229558 4972 scope.go:117] "RemoveContainer" containerID="5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8" Feb 28 11:35:19 crc kubenswrapper[4972]: E0228 11:35:19.229962 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8\": container with ID starting with 5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8 not found: ID does not exist" containerID="5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.229995 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8"} err="failed to get container status \"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8\": rpc error: code = NotFound desc = could not find container \"5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8\": container with ID starting with 5b7efa109873d21473f0d0f04024ebaf674e23e03180399a60194ed102ad67b8 not found: ID does not exist" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.230016 4972 scope.go:117] "RemoveContainer" containerID="a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f" Feb 28 11:35:19 crc kubenswrapper[4972]: E0228 11:35:19.230534 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f\": container with ID starting with a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f not found: ID does not exist" containerID="a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.230555 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f"} err="failed to get container status \"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f\": rpc error: code = NotFound desc = could not find container \"a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f\": container with ID starting with a3c7fdde8606a671274f3b8d14232b099c18c909225ea182dc66bc4ffdbc731f not found: ID does not exist" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.230566 4972 scope.go:117] "RemoveContainer" containerID="972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70" Feb 28 11:35:19 crc kubenswrapper[4972]: E0228 11:35:19.231285 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70\": container with ID starting with 972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70 not found: ID does not exist" containerID="972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.231310 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70"} err="failed to get container status \"972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70\": rpc error: code = NotFound desc = could not find container \"972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70\": container with ID starting with 972129c5d416c56a27771fd699cd2847e2ad24fa17716f1b51680d78908b1a70 not found: ID does not exist" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.803081 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f3fded-90f5-4410-a9d4-700118e399ea" path="/var/lib/kubelet/pods/09f3fded-90f5-4410-a9d4-700118e399ea/volumes" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.804256 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5eb633d-57f6-4a83-b69e-740ee8ac99d6" path="/var/lib/kubelet/pods/a5eb633d-57f6-4a83-b69e-740ee8ac99d6/volumes" Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.963376 4972 generic.go:334] "Generic (PLEG): container finished" podID="873a91c4-7ba8-40ae-8ee1-9024c6e23650" containerID="6ec91b03f0f7f73adfd090fb5bad0cda99ee8126894df4c06ef49bd608465fcb" exitCode=0 Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.963450 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" event={"ID":"873a91c4-7ba8-40ae-8ee1-9024c6e23650","Type":"ContainerDied","Data":"6ec91b03f0f7f73adfd090fb5bad0cda99ee8126894df4c06ef49bd608465fcb"} Feb 28 11:35:19 crc kubenswrapper[4972]: I0228 11:35:19.963499 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" event={"ID":"873a91c4-7ba8-40ae-8ee1-9024c6e23650","Type":"ContainerStarted","Data":"1505de2183b29b9a7005c23634b7976ddc11bd4e05743ddf527598a6f1dbe468"} Feb 28 11:35:20 crc kubenswrapper[4972]: I0228 11:35:20.007886 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-gzg78"] Feb 28 11:35:20 crc kubenswrapper[4972]: I0228 11:35:20.023851 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jmjz5/crc-debug-gzg78"] Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.090735 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.240880 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svzt2\" (UniqueName: \"kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2\") pod \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.240959 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host\") pod \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\" (UID: \"873a91c4-7ba8-40ae-8ee1-9024c6e23650\") " Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.241411 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host" (OuterVolumeSpecName: "host") pod "873a91c4-7ba8-40ae-8ee1-9024c6e23650" (UID: "873a91c4-7ba8-40ae-8ee1-9024c6e23650"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.254498 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2" (OuterVolumeSpecName: "kube-api-access-svzt2") pod "873a91c4-7ba8-40ae-8ee1-9024c6e23650" (UID: "873a91c4-7ba8-40ae-8ee1-9024c6e23650"). InnerVolumeSpecName "kube-api-access-svzt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.343130 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svzt2\" (UniqueName: \"kubernetes.io/projected/873a91c4-7ba8-40ae-8ee1-9024c6e23650-kube-api-access-svzt2\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.343166 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/873a91c4-7ba8-40ae-8ee1-9024c6e23650-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.801316 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="873a91c4-7ba8-40ae-8ee1-9024c6e23650" path="/var/lib/kubelet/pods/873a91c4-7ba8-40ae-8ee1-9024c6e23650/volumes" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.987192 4972 scope.go:117] "RemoveContainer" containerID="6ec91b03f0f7f73adfd090fb5bad0cda99ee8126894df4c06ef49bd608465fcb" Feb 28 11:35:21 crc kubenswrapper[4972]: I0228 11:35:21.987236 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/crc-debug-gzg78" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.415520 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f94fccfd-62888_e50121ef-8724-4a1d-8ee2-badc2c40df58/barbican-api/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.548237 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f94fccfd-62888_e50121ef-8724-4a1d-8ee2-badc2c40df58/barbican-api-log/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.622864 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64d46dc98-5dqwn_fed4da99-54b6-4444-8f0f-4489a5286e4d/barbican-keystone-listener/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.734620 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64d46dc98-5dqwn_fed4da99-54b6-4444-8f0f-4489a5286e4d/barbican-keystone-listener-log/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.807520 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9686d86c-lmjsc_31309c10-9f68-4aed-ab99-dcef67e12b92/barbican-worker-log/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.822666 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9686d86c-lmjsc_31309c10-9f68-4aed-ab99-dcef67e12b92/barbican-worker/0.log" Feb 28 11:35:40 crc kubenswrapper[4972]: I0228 11:35:40.959405 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc_e9d8a507-2de3-4624-ad38-5d50c0c50874/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.055920 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/ceilometer-central-agent/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.143733 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/proxy-httpd/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.183407 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/ceilometer-notification-agent/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.232942 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/sg-core/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.352051 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d0cdd4b-c98a-4297-b44d-9fcb336aecf0/cinder-api-log/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.386661 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d0cdd4b-c98a-4297-b44d-9fcb336aecf0/cinder-api/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.497982 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_84de24a3-0eae-4ec5-af31-a515e98ef012/cinder-scheduler/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.590010 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_84de24a3-0eae-4ec5-af31-a515e98ef012/probe/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.722045 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz_51ff636b-adde-4a49-a96d-bf1275e5b200/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.815834 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-q2clh_5dc25a96-f5f2-491d-b342-41bb674572ba/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:41 crc kubenswrapper[4972]: I0228 11:35:41.905691 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/init/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.109218 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/init/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.118770 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt_c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.195380 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/dnsmasq-dns/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.339216 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e00dd1be-06c1-4c4a-979e-3cb562e7741e/glance-httpd/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.413010 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e00dd1be-06c1-4c4a-979e-3cb562e7741e/glance-log/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.522907 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_35925034-cd2f-4537-8225-0ab87fd3a47e/glance-log/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.561279 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_35925034-cd2f-4537-8225-0ab87fd3a47e/glance-httpd/0.log" Feb 28 11:35:42 crc kubenswrapper[4972]: I0228 11:35:42.679060 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c6cf8f456-722rs_a24fd1fd-763d-4106-87b0-805e2f1fc6c5/horizon/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.006164 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-grxc8_585066f8-afbf-42ac-8163-895ef6b76ec5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.046160 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c6cf8f456-722rs_a24fd1fd-763d-4106-87b0-805e2f1fc6c5/horizon-log/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.069070 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9kfmp_8a963845-f99d-488b-81c8-b13e44cc5ff2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.310628 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29537941-8tjbq_eaa3c5ff-5646-444a-b502-6052100adb2e/keystone-cron/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.376382 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66b556fdbd-nljfl_cdfe3155-40f6-48df-accc-e7667625cd12/keystone-api/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.519592 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_15239450-dea0-4aab-8fe7-1d891d57afab/kube-state-metrics/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.615238 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz_1a18fc65-40ad-474f-b28d-76e3d12cb6df/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.919625 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69dd77478c-x64n9_b7d2ee5d-2b90-4c97-a931-45992f1f50d1/neutron-httpd/0.log" Feb 28 11:35:43 crc kubenswrapper[4972]: I0228 11:35:43.946820 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69dd77478c-x64n9_b7d2ee5d-2b90-4c97-a931-45992f1f50d1/neutron-api/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.020946 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g_ae3ba487-6066-46c6-b9f2-5acc9810d3bf/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.490040 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_64af5818-baad-43d6-abb6-644fb18b3b5e/nova-api-log/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.551779 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3e68715e-2416-4459-8c91-6368dbedb67a/nova-cell0-conductor-conductor/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.769233 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_64af5818-baad-43d6-abb6-644fb18b3b5e/nova-api-api/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.793862 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_3563bc2b-2c86-4824-b238-e8a60f45c60e/nova-cell1-conductor-conductor/0.log" Feb 28 11:35:44 crc kubenswrapper[4972]: I0228 11:35:44.833026 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_70d1c7cf-2d6e-4076-b7cc-8752e5172008/nova-cell1-novncproxy-novncproxy/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.037122 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dqvpd_d0655084-fd3f-4088-bb4e-755a5db445fe/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.219701 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_92b274e8-6674-42a1-81f3-302d28e22926/nova-metadata-log/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.449568 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_44f084cb-b672-497c-8c66-23785c9971f7/nova-scheduler-scheduler/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.492644 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/mysql-bootstrap/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.698573 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/mysql-bootstrap/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.753328 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/galera/0.log" Feb 28 11:35:45 crc kubenswrapper[4972]: I0228 11:35:45.892960 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/mysql-bootstrap/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.098992 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/mysql-bootstrap/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.109926 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/galera/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.276410 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f069c4ff-a971-4849-9d09-4ae37dca70fc/openstackclient/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.300737 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_92b274e8-6674-42a1-81f3-302d28e22926/nova-metadata-metadata/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.377634 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jl8t9_0be6aa95-e5d9-4e85-9991-d58945420d69/openstack-network-exporter/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.554607 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server-init/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.682720 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server-init/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.686230 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovs-vswitchd/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.780372 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server/0.log" Feb 28 11:35:46 crc kubenswrapper[4972]: I0228 11:35:46.909449 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p6jmv_d90126a7-5923-45bd-9e61-29d1e2e3f469/ovn-controller/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.102868 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-k9z8j_d2cb5f2e-f9b1-41fb-8a63-03bebe97019b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.150870 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8ccd0a65-b9b9-4b45-ba11-97903d81f938/openstack-network-exporter/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.228308 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8ccd0a65-b9b9-4b45-ba11-97903d81f938/ovn-northd/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.368020 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6ba1065c-4ef3-473c-9576-d673f19a75a1/openstack-network-exporter/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.405543 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6ba1065c-4ef3-473c-9576-d673f19a75a1/ovsdbserver-nb/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.631064 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec76b1b5-ea80-49cb-ae7c-0190a52640a1/ovsdbserver-sb/0.log" Feb 28 11:35:47 crc kubenswrapper[4972]: I0228 11:35:47.660612 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec76b1b5-ea80-49cb-ae7c-0190a52640a1/openstack-network-exporter/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.118376 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bb67f8988-tt9dr_bae387b2-3007-4c94-9a6f-255bd137d7c0/placement-api/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.137403 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bb67f8988-tt9dr_bae387b2-3007-4c94-9a6f-255bd137d7c0/placement-log/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.212868 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/setup-container/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.436415 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/setup-container/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.436689 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/rabbitmq/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.538558 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/setup-container/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.657804 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/rabbitmq/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.673519 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/setup-container/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.832262 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w_dea3597c-d878-41dd-8952-dda5b8db255e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:48 crc kubenswrapper[4972]: I0228 11:35:48.900089 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-bvts2_a6c60946-e11b-4644-8742-fb1225f8d690/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.015362 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f_a87a4176-9953-45b9-8db1-824b39dd2e4d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.147570 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dskz4_a19c5936-e88b-4336-814d-8b0053fe2247/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.363177 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nqj8m_41b5f77f-8955-4d82-8d40-08c398090bc7/ssh-known-hosts-edpm-deployment/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.568319 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98999f697-8tcfb_09caa13b-95cf-42c1-a35d-7c053e069909/proxy-httpd/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.568678 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2f84x_28171537-6c72-4fe2-9e23-709a2f0824ae/swift-ring-rebalance/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.599812 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98999f697-8tcfb_09caa13b-95cf-42c1-a35d-7c053e069909/proxy-server/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.807111 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-reaper/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.811905 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-auditor/0.log" Feb 28 11:35:49 crc kubenswrapper[4972]: I0228 11:35:49.913355 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-replicator/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.073838 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-server/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.078631 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-auditor/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.096204 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-replicator/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.129319 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-server/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.315992 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-updater/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.326584 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-expirer/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.328136 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-auditor/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.385978 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-replicator/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.516955 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/rsync/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.538649 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-updater/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.547610 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-server/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.618708 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/swift-recon-cron/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.832673 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-shq6f_a667491e-8728-4640-86b9-410f2cf770c9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:50 crc kubenswrapper[4972]: I0228 11:35:50.859306 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_50684e40-2fa2-4cb3-a176-374ab716698c/tempest-tests-tempest-tests-runner/0.log" Feb 28 11:35:51 crc kubenswrapper[4972]: I0228 11:35:51.111599 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz_6436504e-d8a5-4dc9-b500-810a55ef3bbb/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:35:51 crc kubenswrapper[4972]: I0228 11:35:51.126599 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fb460ac4-a796-4700-ab43-565ca0308a20/test-operator-logs-container/0.log" Feb 28 11:35:59 crc kubenswrapper[4972]: I0228 11:35:59.840528 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_56e07ece-d896-4107-b9b1-6f24b64f7985/memcached/0.log" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.141255 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537976-6cw48"] Feb 28 11:36:00 crc kubenswrapper[4972]: E0228 11:36:00.141881 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="873a91c4-7ba8-40ae-8ee1-9024c6e23650" containerName="container-00" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.141897 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="873a91c4-7ba8-40ae-8ee1-9024c6e23650" containerName="container-00" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.142068 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="873a91c4-7ba8-40ae-8ee1-9024c6e23650" containerName="container-00" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.142657 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.144820 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.144970 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.145164 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.149808 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537976-6cw48"] Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.262123 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbjcl\" (UniqueName: \"kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl\") pod \"auto-csr-approver-29537976-6cw48\" (UID: \"b619dc26-f8c5-4808-b31a-fd68eb4d1450\") " pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.364384 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbjcl\" (UniqueName: \"kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl\") pod \"auto-csr-approver-29537976-6cw48\" (UID: \"b619dc26-f8c5-4808-b31a-fd68eb4d1450\") " pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.385087 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbjcl\" (UniqueName: \"kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl\") pod \"auto-csr-approver-29537976-6cw48\" (UID: \"b619dc26-f8c5-4808-b31a-fd68eb4d1450\") " pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.458707 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:00 crc kubenswrapper[4972]: I0228 11:36:00.911257 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537976-6cw48"] Feb 28 11:36:01 crc kubenswrapper[4972]: I0228 11:36:01.386910 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537976-6cw48" event={"ID":"b619dc26-f8c5-4808-b31a-fd68eb4d1450","Type":"ContainerStarted","Data":"c70931d8dc210f3bd96ac0f5e0e9e37ba0c50110bc2b23b32996efe9ab8a3c1c"} Feb 28 11:36:02 crc kubenswrapper[4972]: I0228 11:36:02.397307 4972 generic.go:334] "Generic (PLEG): container finished" podID="b619dc26-f8c5-4808-b31a-fd68eb4d1450" containerID="042679832b747a6e37be2d560ed26a4388ee522e1ef27008a7395e0fe15577a3" exitCode=0 Feb 28 11:36:02 crc kubenswrapper[4972]: I0228 11:36:02.397424 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537976-6cw48" event={"ID":"b619dc26-f8c5-4808-b31a-fd68eb4d1450","Type":"ContainerDied","Data":"042679832b747a6e37be2d560ed26a4388ee522e1ef27008a7395e0fe15577a3"} Feb 28 11:36:03 crc kubenswrapper[4972]: I0228 11:36:03.745368 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:03 crc kubenswrapper[4972]: I0228 11:36:03.832291 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbjcl\" (UniqueName: \"kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl\") pod \"b619dc26-f8c5-4808-b31a-fd68eb4d1450\" (UID: \"b619dc26-f8c5-4808-b31a-fd68eb4d1450\") " Feb 28 11:36:03 crc kubenswrapper[4972]: I0228 11:36:03.838401 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl" (OuterVolumeSpecName: "kube-api-access-gbjcl") pod "b619dc26-f8c5-4808-b31a-fd68eb4d1450" (UID: "b619dc26-f8c5-4808-b31a-fd68eb4d1450"). InnerVolumeSpecName "kube-api-access-gbjcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:36:03 crc kubenswrapper[4972]: I0228 11:36:03.934901 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbjcl\" (UniqueName: \"kubernetes.io/projected/b619dc26-f8c5-4808-b31a-fd68eb4d1450-kube-api-access-gbjcl\") on node \"crc\" DevicePath \"\"" Feb 28 11:36:04 crc kubenswrapper[4972]: I0228 11:36:04.415644 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537976-6cw48" event={"ID":"b619dc26-f8c5-4808-b31a-fd68eb4d1450","Type":"ContainerDied","Data":"c70931d8dc210f3bd96ac0f5e0e9e37ba0c50110bc2b23b32996efe9ab8a3c1c"} Feb 28 11:36:04 crc kubenswrapper[4972]: I0228 11:36:04.415688 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70931d8dc210f3bd96ac0f5e0e9e37ba0c50110bc2b23b32996efe9ab8a3c1c" Feb 28 11:36:04 crc kubenswrapper[4972]: I0228 11:36:04.415740 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537976-6cw48" Feb 28 11:36:04 crc kubenswrapper[4972]: I0228 11:36:04.841301 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537970-7x4x2"] Feb 28 11:36:04 crc kubenswrapper[4972]: I0228 11:36:04.854663 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537970-7x4x2"] Feb 28 11:36:05 crc kubenswrapper[4972]: I0228 11:36:05.802967 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316ae654-63a3-44b0-b947-300ce18544eb" path="/var/lib/kubelet/pods/316ae654-63a3-44b0-b947-300ce18544eb/volumes" Feb 28 11:36:15 crc kubenswrapper[4972]: I0228 11:36:15.360375 4972 scope.go:117] "RemoveContainer" containerID="8f6595d4f6d13a468c6831c50d4c53063d3d9f3bfa278afbfb27e7221c4a7764" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.160895 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.323082 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.359615 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.376004 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.553707 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.566669 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/extract/0.log" Feb 28 11:36:18 crc kubenswrapper[4972]: I0228 11:36:18.591689 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:36:19 crc kubenswrapper[4972]: I0228 11:36:19.030684 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-55cc45767f-krrsg_61ba5500-d464-418c-9248-bc05118b0632/manager/0.log" Feb 28 11:36:19 crc kubenswrapper[4972]: I0228 11:36:19.416579 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5fb5494549-ft8w9_fe59fc2c-7e44-4b09-9d79-415c90f1d3fa/manager/0.log" Feb 28 11:36:19 crc kubenswrapper[4972]: I0228 11:36:19.671704 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-585b788787-j2pnv_3abb4675-640d-4b5c-b3fb-55279630aecd/manager/0.log" Feb 28 11:36:19 crc kubenswrapper[4972]: I0228 11:36:19.919146 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7db95d7ffb-wdlpp_fb530b5d-27e7-4973-9fb5-41b935e9c0a0/manager/0.log" Feb 28 11:36:20 crc kubenswrapper[4972]: I0228 11:36:20.293093 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-768c8b45bb-w987c_7885ab2e-6366-4513-8904-a982144ac4c4/manager/0.log" Feb 28 11:36:20 crc kubenswrapper[4972]: I0228 11:36:20.501017 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-8784b4656-hhmnw_20457d28-02f6-4d31-9938-dbdec910d3c8/manager/0.log" Feb 28 11:36:20 crc kubenswrapper[4972]: I0228 11:36:20.540317 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-c77466965-bkt68_0c55bc45-0607-477d-893f-5782b00cffb2/manager/0.log" Feb 28 11:36:20 crc kubenswrapper[4972]: I0228 11:36:20.801669 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-78b64779b9-xt2pl_15d3191f-413a-4806-8e55-9dd19db4c335/manager/0.log" Feb 28 11:36:20 crc kubenswrapper[4972]: I0228 11:36:20.834218 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6f6f57b9b6-rxzp8_af73edbb-e925-4e12-9cce-58fe2292a497/manager/0.log" Feb 28 11:36:21 crc kubenswrapper[4972]: I0228 11:36:21.067104 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-745fc45789-25bfz_f2d950d1-acba-46f8-99c7-a932569c1c49/manager/0.log" Feb 28 11:36:21 crc kubenswrapper[4972]: I0228 11:36:21.275879 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-768f998cf4-4mxwn_e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86/manager/0.log" Feb 28 11:36:21 crc kubenswrapper[4972]: I0228 11:36:21.486720 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c67ff7674-v8vzt_adaec4ab-a8f3-4719-809d-b102d8e14b7b/manager/0.log" Feb 28 11:36:21 crc kubenswrapper[4972]: I0228 11:36:21.577555 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-cc79fdffd-lgrm6_bae14f0d-c03b-4bf4-9812-2a4cdcae1d94/manager/0.log" Feb 28 11:36:21 crc kubenswrapper[4972]: I0228 11:36:21.726160 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw_473197b6-6493-4509-920b-ab1be61070f9/manager/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.061525 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6cdcc94fb7-qfx4x_cf37191c-db05-439a-9353-1e15f8147289/operator/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.193922 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pvw8x_8e82abb6-3796-42c1-bfe7-688f9117943f/registry-server/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.373074 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-684c7d77b-x7wnv_7068aaa9-0410-48a7-9122-53dfad3b2e67/manager/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.577056 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-bff955cc4-gxsgb_9c3297f2-e6e9-41e0-b299-fae7573d16cc/manager/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.719197 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-8ksfr_5c4dd163-6c95-4f91-8c40-b242a0d191e3/operator/0.log" Feb 28 11:36:22 crc kubenswrapper[4972]: I0228 11:36:22.895131 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-55f4bf89cb-jm2wm_510043b9-9bba-49a8-9902-e37f4564bbe8/manager/0.log" Feb 28 11:36:23 crc kubenswrapper[4972]: I0228 11:36:23.165843 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-56dc67d744-trxff_2da6f8cf-2f39-4614-be52-b6598a919217/manager/0.log" Feb 28 11:36:23 crc kubenswrapper[4972]: I0228 11:36:23.194295 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8467ccb4c8-d9mct_81c08d80-fa23-44c5-9ddf-65dc85b18696/manager/0.log" Feb 28 11:36:23 crc kubenswrapper[4972]: I0228 11:36:23.387732 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cf56d77c9-f7stz_8cce4f5d-78aa-44c4-8c60-8bce681ada4e/manager/0.log" Feb 28 11:36:23 crc kubenswrapper[4972]: I0228 11:36:23.452498 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-65c9f4f6b-vpxcf_65307cfe-55b4-486c-b154-2196bd0a1012/manager/0.log" Feb 28 11:36:25 crc kubenswrapper[4972]: I0228 11:36:25.108134 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6fb74c6d59-x6hp2_9c739be1-15cd-4044-8cce-c3d2750d030e/manager/0.log" Feb 28 11:36:45 crc kubenswrapper[4972]: I0228 11:36:45.272130 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gzmqk_8066e21c-fa0c-497d-8bc8-a27d07dab5b4/control-plane-machine-set-operator/0.log" Feb 28 11:36:45 crc kubenswrapper[4972]: I0228 11:36:45.510418 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xwmd_c9173e02-eddc-40e3-86ab-0f0e8a8f50ac/machine-api-operator/0.log" Feb 28 11:36:45 crc kubenswrapper[4972]: I0228 11:36:45.516780 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xwmd_c9173e02-eddc-40e3-86ab-0f0e8a8f50ac/kube-rbac-proxy/0.log" Feb 28 11:36:58 crc kubenswrapper[4972]: I0228 11:36:58.890285 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:36:58 crc kubenswrapper[4972]: I0228 11:36:58.890847 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.401933 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:36:59 crc kubenswrapper[4972]: E0228 11:36:59.402364 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b619dc26-f8c5-4808-b31a-fd68eb4d1450" containerName="oc" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.402384 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="b619dc26-f8c5-4808-b31a-fd68eb4d1450" containerName="oc" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.402668 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="b619dc26-f8c5-4808-b31a-fd68eb4d1450" containerName="oc" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.404245 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.436558 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.533940 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmn88\" (UniqueName: \"kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.534269 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.534553 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.636237 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmn88\" (UniqueName: \"kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.636320 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.636406 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.636876 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.636898 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.668837 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmn88\" (UniqueName: \"kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88\") pod \"redhat-operators-rlmmn\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:36:59 crc kubenswrapper[4972]: I0228 11:36:59.724190 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.243620 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.251446 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mrgf8_74cd6520-9a2a-4e93-8536-2817d087ba65/cert-manager-controller/0.log" Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.491930 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-ptx2h_0dcc4f20-2ac3-4a75-8674-dfe9800d1844/cert-manager-cainjector/0.log" Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.619754 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-76jg5_c0da0fed-5127-4b89-a79f-73e3d9c1ceee/cert-manager-webhook/0.log" Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.977901 4972 generic.go:334] "Generic (PLEG): container finished" podID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerID="19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e" exitCode=0 Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.977943 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerDied","Data":"19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e"} Feb 28 11:37:00 crc kubenswrapper[4972]: I0228 11:37:00.977969 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerStarted","Data":"ca6cec286e904d1d4f16b5c8174403a5443f716aed9ac79c1ad653012b41922b"} Feb 28 11:37:02 crc kubenswrapper[4972]: I0228 11:37:02.994311 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerStarted","Data":"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1"} Feb 28 11:37:07 crc kubenswrapper[4972]: I0228 11:37:07.030913 4972 generic.go:334] "Generic (PLEG): container finished" podID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerID="7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1" exitCode=0 Feb 28 11:37:07 crc kubenswrapper[4972]: I0228 11:37:07.031370 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerDied","Data":"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1"} Feb 28 11:37:07 crc kubenswrapper[4972]: I0228 11:37:07.034299 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:37:08 crc kubenswrapper[4972]: I0228 11:37:08.043279 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerStarted","Data":"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded"} Feb 28 11:37:08 crc kubenswrapper[4972]: I0228 11:37:08.069300 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rlmmn" podStartSLOduration=2.458437106 podStartE2EDuration="9.069281829s" podCreationTimestamp="2026-02-28 11:36:59 +0000 UTC" firstStartedPulling="2026-02-28 11:37:00.979762532 +0000 UTC m=+3677.891770270" lastFinishedPulling="2026-02-28 11:37:07.590607255 +0000 UTC m=+3684.502614993" observedRunningTime="2026-02-28 11:37:08.062729941 +0000 UTC m=+3684.974737719" watchObservedRunningTime="2026-02-28 11:37:08.069281829 +0000 UTC m=+3684.981289567" Feb 28 11:37:09 crc kubenswrapper[4972]: I0228 11:37:09.724551 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:09 crc kubenswrapper[4972]: I0228 11:37:09.725846 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:10 crc kubenswrapper[4972]: I0228 11:37:10.782408 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rlmmn" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="registry-server" probeResult="failure" output=< Feb 28 11:37:10 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:37:10 crc kubenswrapper[4972]: > Feb 28 11:37:16 crc kubenswrapper[4972]: I0228 11:37:16.907341 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-pdwzw_d5746037-ce0c-44c8-9874-23ac38ad6509/nmstate-console-plugin/0.log" Feb 28 11:37:17 crc kubenswrapper[4972]: I0228 11:37:17.025660 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-tt7f6_1013fa3e-26c1-4b74-a4f5-55c3b18fe34d/nmstate-handler/0.log" Feb 28 11:37:17 crc kubenswrapper[4972]: I0228 11:37:17.078994 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-sx977_a35c5570-d7ee-438e-8768-0daf418f5304/kube-rbac-proxy/0.log" Feb 28 11:37:17 crc kubenswrapper[4972]: I0228 11:37:17.159216 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-sx977_a35c5570-d7ee-438e-8768-0daf418f5304/nmstate-metrics/0.log" Feb 28 11:37:17 crc kubenswrapper[4972]: I0228 11:37:17.296006 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-srsp6_d8bdef5d-8a08-4e53-bb43-2f721ed3570e/nmstate-operator/0.log" Feb 28 11:37:17 crc kubenswrapper[4972]: I0228 11:37:17.469378 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-2z62s_b9a8aae0-4abb-4237-b2da-36542bac6c31/nmstate-webhook/0.log" Feb 28 11:37:19 crc kubenswrapper[4972]: I0228 11:37:19.804923 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:19 crc kubenswrapper[4972]: I0228 11:37:19.883160 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:20 crc kubenswrapper[4972]: I0228 11:37:20.044252 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.149036 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rlmmn" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="registry-server" containerID="cri-o://ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded" gracePeriod=2 Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.632087 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.669150 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content\") pod \"5797c733-fdb6-4ffd-b1b7-2952daad3406\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.669274 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmn88\" (UniqueName: \"kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88\") pod \"5797c733-fdb6-4ffd-b1b7-2952daad3406\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.669367 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities\") pod \"5797c733-fdb6-4ffd-b1b7-2952daad3406\" (UID: \"5797c733-fdb6-4ffd-b1b7-2952daad3406\") " Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.669840 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities" (OuterVolumeSpecName: "utilities") pod "5797c733-fdb6-4ffd-b1b7-2952daad3406" (UID: "5797c733-fdb6-4ffd-b1b7-2952daad3406"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.670282 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.677691 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88" (OuterVolumeSpecName: "kube-api-access-fmn88") pod "5797c733-fdb6-4ffd-b1b7-2952daad3406" (UID: "5797c733-fdb6-4ffd-b1b7-2952daad3406"). InnerVolumeSpecName "kube-api-access-fmn88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.768829 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5797c733-fdb6-4ffd-b1b7-2952daad3406" (UID: "5797c733-fdb6-4ffd-b1b7-2952daad3406"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.771606 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5797c733-fdb6-4ffd-b1b7-2952daad3406-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:37:21 crc kubenswrapper[4972]: I0228 11:37:21.771635 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmn88\" (UniqueName: \"kubernetes.io/projected/5797c733-fdb6-4ffd-b1b7-2952daad3406-kube-api-access-fmn88\") on node \"crc\" DevicePath \"\"" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.159040 4972 generic.go:334] "Generic (PLEG): container finished" podID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerID="ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded" exitCode=0 Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.159094 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerDied","Data":"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded"} Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.159126 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rlmmn" event={"ID":"5797c733-fdb6-4ffd-b1b7-2952daad3406","Type":"ContainerDied","Data":"ca6cec286e904d1d4f16b5c8174403a5443f716aed9ac79c1ad653012b41922b"} Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.159149 4972 scope.go:117] "RemoveContainer" containerID="ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.159289 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rlmmn" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.184086 4972 scope.go:117] "RemoveContainer" containerID="7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.193582 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.209635 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rlmmn"] Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.221616 4972 scope.go:117] "RemoveContainer" containerID="19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.274140 4972 scope.go:117] "RemoveContainer" containerID="ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded" Feb 28 11:37:22 crc kubenswrapper[4972]: E0228 11:37:22.274511 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded\": container with ID starting with ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded not found: ID does not exist" containerID="ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.274545 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded"} err="failed to get container status \"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded\": rpc error: code = NotFound desc = could not find container \"ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded\": container with ID starting with ad69ad378cc2556413dde5cff42c88c8bdd26a2be8c8cf4ed3f3a1d10f623ded not found: ID does not exist" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.274566 4972 scope.go:117] "RemoveContainer" containerID="7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1" Feb 28 11:37:22 crc kubenswrapper[4972]: E0228 11:37:22.277812 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1\": container with ID starting with 7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1 not found: ID does not exist" containerID="7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.277860 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1"} err="failed to get container status \"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1\": rpc error: code = NotFound desc = could not find container \"7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1\": container with ID starting with 7f373db75f22485236817f9b5f215644d5d11738466de0f75a8613a0e93e4ba1 not found: ID does not exist" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.277889 4972 scope.go:117] "RemoveContainer" containerID="19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e" Feb 28 11:37:22 crc kubenswrapper[4972]: E0228 11:37:22.278185 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e\": container with ID starting with 19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e not found: ID does not exist" containerID="19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e" Feb 28 11:37:22 crc kubenswrapper[4972]: I0228 11:37:22.278238 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e"} err="failed to get container status \"19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e\": rpc error: code = NotFound desc = could not find container \"19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e\": container with ID starting with 19fc97976f249902699e2a9a22c5f28f92c26c87e4a76205e685faf0ed8d779e not found: ID does not exist" Feb 28 11:37:23 crc kubenswrapper[4972]: I0228 11:37:23.800798 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" path="/var/lib/kubelet/pods/5797c733-fdb6-4ffd-b1b7-2952daad3406/volumes" Feb 28 11:37:28 crc kubenswrapper[4972]: I0228 11:37:28.890574 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:37:28 crc kubenswrapper[4972]: I0228 11:37:28.890932 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:37:47 crc kubenswrapper[4972]: I0228 11:37:47.980802 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-lxdv7_47d6aa51-c3aa-4f21-a02f-36663c003b7f/kube-rbac-proxy/0.log" Feb 28 11:37:47 crc kubenswrapper[4972]: I0228 11:37:47.981608 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-lxdv7_47d6aa51-c3aa-4f21-a02f-36663c003b7f/controller/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.130427 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.328501 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.355040 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.365993 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.373441 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.534275 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.552845 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.563180 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.592483 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.789290 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.792419 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.796730 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/controller/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.813510 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:37:48 crc kubenswrapper[4972]: I0228 11:37:48.998951 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/kube-rbac-proxy/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.013449 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/frr-metrics/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.016754 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/kube-rbac-proxy-frr/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.234056 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-c9c7m_c4ddb606-f9be-489d-8c0b-5c89adbb265a/frr-k8s-webhook-server/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.262907 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/reloader/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.463255 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57bd854678-2r552_7d4fb02a-0341-49cf-947f-86fb93ac710b/manager/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.601388 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5cdfc6748f-z4vkk_ac415cf4-e2bc-471a-91aa-2752b88b2636/webhook-server/0.log" Feb 28 11:37:49 crc kubenswrapper[4972]: I0228 11:37:49.749315 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n2xm6_b387ac1b-669a-43f3-b336-7df69f886a1a/kube-rbac-proxy/0.log" Feb 28 11:37:50 crc kubenswrapper[4972]: I0228 11:37:50.281068 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n2xm6_b387ac1b-669a-43f3-b336-7df69f886a1a/speaker/0.log" Feb 28 11:37:50 crc kubenswrapper[4972]: I0228 11:37:50.524308 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/frr/0.log" Feb 28 11:37:58 crc kubenswrapper[4972]: I0228 11:37:58.891240 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:37:58 crc kubenswrapper[4972]: I0228 11:37:58.892085 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:37:58 crc kubenswrapper[4972]: I0228 11:37:58.892157 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:37:58 crc kubenswrapper[4972]: I0228 11:37:58.893280 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:37:58 crc kubenswrapper[4972]: I0228 11:37:58.893380 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7" gracePeriod=600 Feb 28 11:37:59 crc kubenswrapper[4972]: I0228 11:37:59.539078 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7" exitCode=0 Feb 28 11:37:59 crc kubenswrapper[4972]: I0228 11:37:59.539170 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7"} Feb 28 11:37:59 crc kubenswrapper[4972]: I0228 11:37:59.539570 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3"} Feb 28 11:37:59 crc kubenswrapper[4972]: I0228 11:37:59.539594 4972 scope.go:117] "RemoveContainer" containerID="fbb4808c491609916e908a27cd759647d82ae00be3ca4a1697d7d9f1f920938e" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.184526 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537978-42wm9"] Feb 28 11:38:00 crc kubenswrapper[4972]: E0228 11:38:00.185708 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="extract-utilities" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.185733 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="extract-utilities" Feb 28 11:38:00 crc kubenswrapper[4972]: E0228 11:38:00.185774 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="registry-server" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.185786 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="registry-server" Feb 28 11:38:00 crc kubenswrapper[4972]: E0228 11:38:00.185831 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="extract-content" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.185844 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="extract-content" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.186182 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5797c733-fdb6-4ffd-b1b7-2952daad3406" containerName="registry-server" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.187296 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.190966 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537978-42wm9"] Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.192010 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.192151 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.192322 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.266252 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfdvk\" (UniqueName: \"kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk\") pod \"auto-csr-approver-29537978-42wm9\" (UID: \"4142236e-468b-4859-b322-4899a912363f\") " pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.368588 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfdvk\" (UniqueName: \"kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk\") pod \"auto-csr-approver-29537978-42wm9\" (UID: \"4142236e-468b-4859-b322-4899a912363f\") " pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.397114 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfdvk\" (UniqueName: \"kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk\") pod \"auto-csr-approver-29537978-42wm9\" (UID: \"4142236e-468b-4859-b322-4899a912363f\") " pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.526070 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:00 crc kubenswrapper[4972]: I0228 11:38:00.827292 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537978-42wm9"] Feb 28 11:38:01 crc kubenswrapper[4972]: I0228 11:38:01.575816 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537978-42wm9" event={"ID":"4142236e-468b-4859-b322-4899a912363f","Type":"ContainerStarted","Data":"ef2904ca2fc7de37722da6d163f127463520adf2037acb2d13b5197974236ce2"} Feb 28 11:38:02 crc kubenswrapper[4972]: I0228 11:38:02.589724 4972 generic.go:334] "Generic (PLEG): container finished" podID="4142236e-468b-4859-b322-4899a912363f" containerID="78d2c2e3d0cc3e115c09ffd2b8416df25ea52397c018e2032d7b0e66ac28a300" exitCode=0 Feb 28 11:38:02 crc kubenswrapper[4972]: I0228 11:38:02.589975 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537978-42wm9" event={"ID":"4142236e-468b-4859-b322-4899a912363f","Type":"ContainerDied","Data":"78d2c2e3d0cc3e115c09ffd2b8416df25ea52397c018e2032d7b0e66ac28a300"} Feb 28 11:38:03 crc kubenswrapper[4972]: I0228 11:38:03.986357 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.078258 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfdvk\" (UniqueName: \"kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk\") pod \"4142236e-468b-4859-b322-4899a912363f\" (UID: \"4142236e-468b-4859-b322-4899a912363f\") " Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.088428 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk" (OuterVolumeSpecName: "kube-api-access-bfdvk") pod "4142236e-468b-4859-b322-4899a912363f" (UID: "4142236e-468b-4859-b322-4899a912363f"). InnerVolumeSpecName "kube-api-access-bfdvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.181330 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfdvk\" (UniqueName: \"kubernetes.io/projected/4142236e-468b-4859-b322-4899a912363f-kube-api-access-bfdvk\") on node \"crc\" DevicePath \"\"" Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.612153 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537978-42wm9" event={"ID":"4142236e-468b-4859-b322-4899a912363f","Type":"ContainerDied","Data":"ef2904ca2fc7de37722da6d163f127463520adf2037acb2d13b5197974236ce2"} Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.612221 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef2904ca2fc7de37722da6d163f127463520adf2037acb2d13b5197974236ce2" Feb 28 11:38:04 crc kubenswrapper[4972]: I0228 11:38:04.612224 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537978-42wm9" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.057701 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537972-t9k8k"] Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.067195 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537972-t9k8k"] Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.292068 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.454987 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.460920 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.460955 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.675571 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.687092 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.761499 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/extract/0.log" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.800608 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c4beba-180c-45c5-9012-f3a483a9844b" path="/var/lib/kubelet/pods/e3c4beba-180c-45c5-9012-f3a483a9844b/volumes" Feb 28 11:38:05 crc kubenswrapper[4972]: I0228 11:38:05.863407 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.070703 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.117331 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.146788 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.263178 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.298770 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.489643 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.721059 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/registry-server/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.775018 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.794155 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.800631 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.972382 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:38:06 crc kubenswrapper[4972]: I0228 11:38:06.988680 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.228354 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.435810 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.447739 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/registry-server/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.497504 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.523507 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.637434 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.658149 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.675861 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/extract/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.898052 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:38:07 crc kubenswrapper[4972]: I0228 11:38:07.929911 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-m88zk_72e5fba8-6a6e-4310-bcb6-ad064ec74bb9/marketplace-operator/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.286231 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.314180 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.334451 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.506820 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.543264 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.694109 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/registry-server/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.713503 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.919638 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.946796 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:38:08 crc kubenswrapper[4972]: I0228 11:38:08.953720 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:38:09 crc kubenswrapper[4972]: I0228 11:38:09.189705 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:38:09 crc kubenswrapper[4972]: I0228 11:38:09.189772 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:38:09 crc kubenswrapper[4972]: I0228 11:38:09.767425 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/registry-server/0.log" Feb 28 11:38:15 crc kubenswrapper[4972]: I0228 11:38:15.528102 4972 scope.go:117] "RemoveContainer" containerID="56d84b1356b7569072a8aa9c55752e41664847d1f01e7fd637010b8660455d1e" Feb 28 11:38:43 crc kubenswrapper[4972]: E0228 11:38:43.364654 4972 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.12:60434->38.102.83.12:44997: read tcp 38.102.83.12:60434->38.102.83.12:44997: read: connection reset by peer Feb 28 11:39:07 crc kubenswrapper[4972]: I0228 11:39:07.716109 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="44e7c813-beaa-4015-a581-afcaea85077c" containerName="galera" probeResult="failure" output="command timed out" Feb 28 11:39:53 crc kubenswrapper[4972]: I0228 11:39:53.932497 4972 generic.go:334] "Generic (PLEG): container finished" podID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerID="14e22614dc3ff190e17c999c58a137372f7236e13b44de008bd27336ef3f4846" exitCode=0 Feb 28 11:39:53 crc kubenswrapper[4972]: I0228 11:39:53.932618 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jmjz5/must-gather-s544x" event={"ID":"c9b9a51d-5064-4502-a8fc-3a1cc75d591b","Type":"ContainerDied","Data":"14e22614dc3ff190e17c999c58a137372f7236e13b44de008bd27336ef3f4846"} Feb 28 11:39:53 crc kubenswrapper[4972]: I0228 11:39:53.934840 4972 scope.go:117] "RemoveContainer" containerID="14e22614dc3ff190e17c999c58a137372f7236e13b44de008bd27336ef3f4846" Feb 28 11:39:54 crc kubenswrapper[4972]: I0228 11:39:54.669002 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jmjz5_must-gather-s544x_c9b9a51d-5064-4502-a8fc-3a1cc75d591b/gather/0.log" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.162670 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537980-qtngz"] Feb 28 11:40:00 crc kubenswrapper[4972]: E0228 11:40:00.163945 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4142236e-468b-4859-b322-4899a912363f" containerName="oc" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.163964 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4142236e-468b-4859-b322-4899a912363f" containerName="oc" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.164188 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4142236e-468b-4859-b322-4899a912363f" containerName="oc" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.164930 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.167612 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.168292 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.168851 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.194816 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537980-qtngz"] Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.283426 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwfrl\" (UniqueName: \"kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl\") pod \"auto-csr-approver-29537980-qtngz\" (UID: \"2f0da5c8-255b-46b6-acf9-1cd155990486\") " pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.385599 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwfrl\" (UniqueName: \"kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl\") pod \"auto-csr-approver-29537980-qtngz\" (UID: \"2f0da5c8-255b-46b6-acf9-1cd155990486\") " pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.416162 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwfrl\" (UniqueName: \"kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl\") pod \"auto-csr-approver-29537980-qtngz\" (UID: \"2f0da5c8-255b-46b6-acf9-1cd155990486\") " pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:00 crc kubenswrapper[4972]: I0228 11:40:00.498438 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:01 crc kubenswrapper[4972]: I0228 11:40:01.053901 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537980-qtngz"] Feb 28 11:40:02 crc kubenswrapper[4972]: I0228 11:40:02.025234 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537980-qtngz" event={"ID":"2f0da5c8-255b-46b6-acf9-1cd155990486","Type":"ContainerStarted","Data":"1ed88ed92c20cd76799be47ed378a9bb82733f6b1394d43b6d29e1fbc4d3fabf"} Feb 28 11:40:02 crc kubenswrapper[4972]: I0228 11:40:02.520243 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jmjz5/must-gather-s544x"] Feb 28 11:40:02 crc kubenswrapper[4972]: I0228 11:40:02.520938 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jmjz5/must-gather-s544x" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="copy" containerID="cri-o://e8e5d4c2262051ebb4fba44c6bb5c34add63d6c9f1f0482324754e65cb9f1abd" gracePeriod=2 Feb 28 11:40:02 crc kubenswrapper[4972]: I0228 11:40:02.531790 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jmjz5/must-gather-s544x"] Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.035189 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537980-qtngz" event={"ID":"2f0da5c8-255b-46b6-acf9-1cd155990486","Type":"ContainerStarted","Data":"5cba31be8163144c6fc906581293aa0166e437b5af21f5829f824c195ce499ba"} Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.042270 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jmjz5_must-gather-s544x_c9b9a51d-5064-4502-a8fc-3a1cc75d591b/copy/0.log" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.042613 4972 generic.go:334] "Generic (PLEG): container finished" podID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerID="e8e5d4c2262051ebb4fba44c6bb5c34add63d6c9f1f0482324754e65cb9f1abd" exitCode=143 Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.042648 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f70d1ff76802cce6b0702778dff204c6033c993eeb49abec15a1a66b6a02214" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.052718 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537980-qtngz" podStartSLOduration=1.7994260579999999 podStartE2EDuration="3.05269932s" podCreationTimestamp="2026-02-28 11:40:00 +0000 UTC" firstStartedPulling="2026-02-28 11:40:01.058655383 +0000 UTC m=+3857.970663121" lastFinishedPulling="2026-02-28 11:40:02.311928605 +0000 UTC m=+3859.223936383" observedRunningTime="2026-02-28 11:40:03.05124684 +0000 UTC m=+3859.963254578" watchObservedRunningTime="2026-02-28 11:40:03.05269932 +0000 UTC m=+3859.964707058" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.072732 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jmjz5_must-gather-s544x_c9b9a51d-5064-4502-a8fc-3a1cc75d591b/copy/0.log" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.073096 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.148090 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output\") pod \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.148379 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrzvp\" (UniqueName: \"kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp\") pod \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\" (UID: \"c9b9a51d-5064-4502-a8fc-3a1cc75d591b\") " Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.153713 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp" (OuterVolumeSpecName: "kube-api-access-zrzvp") pod "c9b9a51d-5064-4502-a8fc-3a1cc75d591b" (UID: "c9b9a51d-5064-4502-a8fc-3a1cc75d591b"). InnerVolumeSpecName "kube-api-access-zrzvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.252300 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrzvp\" (UniqueName: \"kubernetes.io/projected/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-kube-api-access-zrzvp\") on node \"crc\" DevicePath \"\"" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.303144 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c9b9a51d-5064-4502-a8fc-3a1cc75d591b" (UID: "c9b9a51d-5064-4502-a8fc-3a1cc75d591b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.355237 4972 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c9b9a51d-5064-4502-a8fc-3a1cc75d591b-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 28 11:40:03 crc kubenswrapper[4972]: I0228 11:40:03.803236 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" path="/var/lib/kubelet/pods/c9b9a51d-5064-4502-a8fc-3a1cc75d591b/volumes" Feb 28 11:40:04 crc kubenswrapper[4972]: I0228 11:40:04.054616 4972 generic.go:334] "Generic (PLEG): container finished" podID="2f0da5c8-255b-46b6-acf9-1cd155990486" containerID="5cba31be8163144c6fc906581293aa0166e437b5af21f5829f824c195ce499ba" exitCode=0 Feb 28 11:40:04 crc kubenswrapper[4972]: I0228 11:40:04.054701 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537980-qtngz" event={"ID":"2f0da5c8-255b-46b6-acf9-1cd155990486","Type":"ContainerDied","Data":"5cba31be8163144c6fc906581293aa0166e437b5af21f5829f824c195ce499ba"} Feb 28 11:40:04 crc kubenswrapper[4972]: I0228 11:40:04.054724 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jmjz5/must-gather-s544x" Feb 28 11:40:05 crc kubenswrapper[4972]: I0228 11:40:05.483774 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:05 crc kubenswrapper[4972]: I0228 11:40:05.637500 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwfrl\" (UniqueName: \"kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl\") pod \"2f0da5c8-255b-46b6-acf9-1cd155990486\" (UID: \"2f0da5c8-255b-46b6-acf9-1cd155990486\") " Feb 28 11:40:05 crc kubenswrapper[4972]: I0228 11:40:05.646853 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl" (OuterVolumeSpecName: "kube-api-access-bwfrl") pod "2f0da5c8-255b-46b6-acf9-1cd155990486" (UID: "2f0da5c8-255b-46b6-acf9-1cd155990486"). InnerVolumeSpecName "kube-api-access-bwfrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:40:05 crc kubenswrapper[4972]: I0228 11:40:05.740828 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwfrl\" (UniqueName: \"kubernetes.io/projected/2f0da5c8-255b-46b6-acf9-1cd155990486-kube-api-access-bwfrl\") on node \"crc\" DevicePath \"\"" Feb 28 11:40:06 crc kubenswrapper[4972]: I0228 11:40:06.079698 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537980-qtngz" event={"ID":"2f0da5c8-255b-46b6-acf9-1cd155990486","Type":"ContainerDied","Data":"1ed88ed92c20cd76799be47ed378a9bb82733f6b1394d43b6d29e1fbc4d3fabf"} Feb 28 11:40:06 crc kubenswrapper[4972]: I0228 11:40:06.080014 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed88ed92c20cd76799be47ed378a9bb82733f6b1394d43b6d29e1fbc4d3fabf" Feb 28 11:40:06 crc kubenswrapper[4972]: I0228 11:40:06.079746 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537980-qtngz" Feb 28 11:40:06 crc kubenswrapper[4972]: I0228 11:40:06.138344 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537974-pr6fq"] Feb 28 11:40:06 crc kubenswrapper[4972]: I0228 11:40:06.158375 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537974-pr6fq"] Feb 28 11:40:07 crc kubenswrapper[4972]: I0228 11:40:07.807052 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31243ab6-69bc-46b0-83b9-da39f9f11074" path="/var/lib/kubelet/pods/31243ab6-69bc-46b0-83b9-da39f9f11074/volumes" Feb 28 11:40:15 crc kubenswrapper[4972]: I0228 11:40:15.676119 4972 scope.go:117] "RemoveContainer" containerID="14e22614dc3ff190e17c999c58a137372f7236e13b44de008bd27336ef3f4846" Feb 28 11:40:15 crc kubenswrapper[4972]: I0228 11:40:15.777628 4972 scope.go:117] "RemoveContainer" containerID="9204d1f38c271e8afe07636ec07bfb9355ae86603b6561193a9ef611185b0245" Feb 28 11:40:28 crc kubenswrapper[4972]: I0228 11:40:28.891660 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:40:28 crc kubenswrapper[4972]: I0228 11:40:28.892873 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:40:58 crc kubenswrapper[4972]: I0228 11:40:58.891143 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:40:58 crc kubenswrapper[4972]: I0228 11:40:58.891826 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:41:15 crc kubenswrapper[4972]: I0228 11:41:15.882840 4972 scope.go:117] "RemoveContainer" containerID="e8e5d4c2262051ebb4fba44c6bb5c34add63d6c9f1f0482324754e65cb9f1abd" Feb 28 11:41:15 crc kubenswrapper[4972]: I0228 11:41:15.937321 4972 scope.go:117] "RemoveContainer" containerID="77a679720e8485b9d9503de24014051b6f1912c02b5a8740ae8962f1ae60aa6b" Feb 28 11:41:28 crc kubenswrapper[4972]: I0228 11:41:28.890960 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:41:28 crc kubenswrapper[4972]: I0228 11:41:28.891757 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:41:28 crc kubenswrapper[4972]: I0228 11:41:28.891827 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:41:28 crc kubenswrapper[4972]: I0228 11:41:28.892968 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:41:28 crc kubenswrapper[4972]: I0228 11:41:28.893069 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" gracePeriod=600 Feb 28 11:41:29 crc kubenswrapper[4972]: E0228 11:41:29.025552 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:41:29 crc kubenswrapper[4972]: I0228 11:41:29.057070 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" exitCode=0 Feb 28 11:41:29 crc kubenswrapper[4972]: I0228 11:41:29.057143 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3"} Feb 28 11:41:29 crc kubenswrapper[4972]: I0228 11:41:29.057498 4972 scope.go:117] "RemoveContainer" containerID="28a6c7532889d3dbd5109f99f81b91e2fdd8021aa6f65a29ceae643896ea10a7" Feb 28 11:41:29 crc kubenswrapper[4972]: I0228 11:41:29.058173 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:41:29 crc kubenswrapper[4972]: E0228 11:41:29.058516 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.685221 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:39 crc kubenswrapper[4972]: E0228 11:41:39.687768 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="copy" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.687794 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="copy" Feb 28 11:41:39 crc kubenswrapper[4972]: E0228 11:41:39.687840 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="gather" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.687853 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="gather" Feb 28 11:41:39 crc kubenswrapper[4972]: E0228 11:41:39.687900 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0da5c8-255b-46b6-acf9-1cd155990486" containerName="oc" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.687915 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0da5c8-255b-46b6-acf9-1cd155990486" containerName="oc" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.688313 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0da5c8-255b-46b6-acf9-1cd155990486" containerName="oc" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.688371 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="copy" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.688398 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b9a51d-5064-4502-a8fc-3a1cc75d591b" containerName="gather" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.695305 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.711699 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.789913 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:41:39 crc kubenswrapper[4972]: E0228 11:41:39.790412 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.832369 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.832471 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.832617 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874hg\" (UniqueName: \"kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.934706 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.934932 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874hg\" (UniqueName: \"kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.934979 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.935579 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.935704 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:39 crc kubenswrapper[4972]: I0228 11:41:39.958587 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874hg\" (UniqueName: \"kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg\") pod \"certified-operators-fpxf9\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.063324 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.590061 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.662603 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.664283 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.676314 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.749104 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.749157 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttdgt\" (UniqueName: \"kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.749255 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.851096 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.851205 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.851232 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttdgt\" (UniqueName: \"kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.851576 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.851653 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:40 crc kubenswrapper[4972]: I0228 11:41:40.869491 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttdgt\" (UniqueName: \"kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt\") pod \"community-operators-mlq6d\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:41 crc kubenswrapper[4972]: I0228 11:41:41.017205 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:41 crc kubenswrapper[4972]: I0228 11:41:41.211992 4972 generic.go:334] "Generic (PLEG): container finished" podID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerID="c13f41c2760399b532406fa221416057a3ba766e674f83ad21483f6d6184a120" exitCode=0 Feb 28 11:41:41 crc kubenswrapper[4972]: I0228 11:41:41.212045 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerDied","Data":"c13f41c2760399b532406fa221416057a3ba766e674f83ad21483f6d6184a120"} Feb 28 11:41:41 crc kubenswrapper[4972]: I0228 11:41:41.212115 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerStarted","Data":"478c86b1c70b68ead770ae085c3a3ce89a316c28bd33f3fb50f471ca2747259c"} Feb 28 11:41:41 crc kubenswrapper[4972]: I0228 11:41:41.524035 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:42 crc kubenswrapper[4972]: I0228 11:41:42.224802 4972 generic.go:334] "Generic (PLEG): container finished" podID="243370e7-d511-4d9e-9893-32e6617a9512" containerID="a93b099f73e82ec1fe1e6fdebebd8a90f6447d52e3cfe49398d181bd22c04cc0" exitCode=0 Feb 28 11:41:42 crc kubenswrapper[4972]: I0228 11:41:42.225504 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerDied","Data":"a93b099f73e82ec1fe1e6fdebebd8a90f6447d52e3cfe49398d181bd22c04cc0"} Feb 28 11:41:42 crc kubenswrapper[4972]: I0228 11:41:42.225537 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerStarted","Data":"ed192b87f63482fdb288ea20a816ed5561088529e114841ce730188d9a974903"} Feb 28 11:41:42 crc kubenswrapper[4972]: I0228 11:41:42.231852 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerStarted","Data":"12b1774889766eba58469b4b10c2aed7f225db07fe4e6a376bf8c9d02b09471f"} Feb 28 11:41:43 crc kubenswrapper[4972]: I0228 11:41:43.242472 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerStarted","Data":"35071658ab0e01b7e94b4c66b86bf35d103758c902ac157568aa846cdc97c0c6"} Feb 28 11:41:43 crc kubenswrapper[4972]: I0228 11:41:43.245752 4972 generic.go:334] "Generic (PLEG): container finished" podID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerID="12b1774889766eba58469b4b10c2aed7f225db07fe4e6a376bf8c9d02b09471f" exitCode=0 Feb 28 11:41:43 crc kubenswrapper[4972]: I0228 11:41:43.245781 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerDied","Data":"12b1774889766eba58469b4b10c2aed7f225db07fe4e6a376bf8c9d02b09471f"} Feb 28 11:41:44 crc kubenswrapper[4972]: I0228 11:41:44.268178 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerStarted","Data":"7241fcc9c2b7d1a6c4304e55de2ee25732371a5e5762d1788853994607762be4"} Feb 28 11:41:44 crc kubenswrapper[4972]: I0228 11:41:44.289554 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fpxf9" podStartSLOduration=2.815944637 podStartE2EDuration="5.28953483s" podCreationTimestamp="2026-02-28 11:41:39 +0000 UTC" firstStartedPulling="2026-02-28 11:41:41.214801602 +0000 UTC m=+3958.126809340" lastFinishedPulling="2026-02-28 11:41:43.688391765 +0000 UTC m=+3960.600399533" observedRunningTime="2026-02-28 11:41:44.285650096 +0000 UTC m=+3961.197657884" watchObservedRunningTime="2026-02-28 11:41:44.28953483 +0000 UTC m=+3961.201542558" Feb 28 11:41:45 crc kubenswrapper[4972]: I0228 11:41:45.284388 4972 generic.go:334] "Generic (PLEG): container finished" podID="243370e7-d511-4d9e-9893-32e6617a9512" containerID="35071658ab0e01b7e94b4c66b86bf35d103758c902ac157568aa846cdc97c0c6" exitCode=0 Feb 28 11:41:45 crc kubenswrapper[4972]: I0228 11:41:45.284496 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerDied","Data":"35071658ab0e01b7e94b4c66b86bf35d103758c902ac157568aa846cdc97c0c6"} Feb 28 11:41:46 crc kubenswrapper[4972]: I0228 11:41:46.297785 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerStarted","Data":"fa934fa9850e6749746907a61876adfd513d246fd13d65b7bae878cc40de4b86"} Feb 28 11:41:46 crc kubenswrapper[4972]: I0228 11:41:46.327710 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mlq6d" podStartSLOduration=2.901121669 podStartE2EDuration="6.327684711s" podCreationTimestamp="2026-02-28 11:41:40 +0000 UTC" firstStartedPulling="2026-02-28 11:41:42.226966162 +0000 UTC m=+3959.138973940" lastFinishedPulling="2026-02-28 11:41:45.653529234 +0000 UTC m=+3962.565536982" observedRunningTime="2026-02-28 11:41:46.320174819 +0000 UTC m=+3963.232182567" watchObservedRunningTime="2026-02-28 11:41:46.327684711 +0000 UTC m=+3963.239692479" Feb 28 11:41:50 crc kubenswrapper[4972]: I0228 11:41:50.064298 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:50 crc kubenswrapper[4972]: I0228 11:41:50.065162 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:50 crc kubenswrapper[4972]: I0228 11:41:50.144078 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:50 crc kubenswrapper[4972]: I0228 11:41:50.425614 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:51 crc kubenswrapper[4972]: I0228 11:41:51.017861 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:51 crc kubenswrapper[4972]: I0228 11:41:51.017915 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:51 crc kubenswrapper[4972]: I0228 11:41:51.097708 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:51 crc kubenswrapper[4972]: I0228 11:41:51.398727 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:51 crc kubenswrapper[4972]: I0228 11:41:51.789100 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:41:51 crc kubenswrapper[4972]: E0228 11:41:51.789361 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:41:52 crc kubenswrapper[4972]: I0228 11:41:52.462868 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:52 crc kubenswrapper[4972]: I0228 11:41:52.463227 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fpxf9" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="registry-server" containerID="cri-o://7241fcc9c2b7d1a6c4304e55de2ee25732371a5e5762d1788853994607762be4" gracePeriod=2 Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.419904 4972 generic.go:334] "Generic (PLEG): container finished" podID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerID="7241fcc9c2b7d1a6c4304e55de2ee25732371a5e5762d1788853994607762be4" exitCode=0 Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.420226 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerDied","Data":"7241fcc9c2b7d1a6c4304e55de2ee25732371a5e5762d1788853994607762be4"} Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.468667 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.468929 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mlq6d" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="registry-server" containerID="cri-o://fa934fa9850e6749746907a61876adfd513d246fd13d65b7bae878cc40de4b86" gracePeriod=2 Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.625780 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.758220 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities\") pod \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.758432 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content\") pod \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.758540 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-874hg\" (UniqueName: \"kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg\") pod \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\" (UID: \"239a1f22-e163-4fa8-b9f0-0acc4d729fba\") " Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.760135 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities" (OuterVolumeSpecName: "utilities") pod "239a1f22-e163-4fa8-b9f0-0acc4d729fba" (UID: "239a1f22-e163-4fa8-b9f0-0acc4d729fba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.765668 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg" (OuterVolumeSpecName: "kube-api-access-874hg") pod "239a1f22-e163-4fa8-b9f0-0acc4d729fba" (UID: "239a1f22-e163-4fa8-b9f0-0acc4d729fba"). InnerVolumeSpecName "kube-api-access-874hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.825257 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239a1f22-e163-4fa8-b9f0-0acc4d729fba" (UID: "239a1f22-e163-4fa8-b9f0-0acc4d729fba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.861134 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.861176 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-874hg\" (UniqueName: \"kubernetes.io/projected/239a1f22-e163-4fa8-b9f0-0acc4d729fba-kube-api-access-874hg\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:53 crc kubenswrapper[4972]: I0228 11:41:53.861189 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239a1f22-e163-4fa8-b9f0-0acc4d729fba-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.439088 4972 generic.go:334] "Generic (PLEG): container finished" podID="243370e7-d511-4d9e-9893-32e6617a9512" containerID="fa934fa9850e6749746907a61876adfd513d246fd13d65b7bae878cc40de4b86" exitCode=0 Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.439179 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerDied","Data":"fa934fa9850e6749746907a61876adfd513d246fd13d65b7bae878cc40de4b86"} Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.442636 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpxf9" event={"ID":"239a1f22-e163-4fa8-b9f0-0acc4d729fba","Type":"ContainerDied","Data":"478c86b1c70b68ead770ae085c3a3ce89a316c28bd33f3fb50f471ca2747259c"} Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.442681 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpxf9" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.442716 4972 scope.go:117] "RemoveContainer" containerID="7241fcc9c2b7d1a6c4304e55de2ee25732371a5e5762d1788853994607762be4" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.465446 4972 scope.go:117] "RemoveContainer" containerID="12b1774889766eba58469b4b10c2aed7f225db07fe4e6a376bf8c9d02b09471f" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.485009 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.496867 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fpxf9"] Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.510782 4972 scope.go:117] "RemoveContainer" containerID="c13f41c2760399b532406fa221416057a3ba766e674f83ad21483f6d6184a120" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.605965 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.785753 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities\") pod \"243370e7-d511-4d9e-9893-32e6617a9512\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.785878 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content\") pod \"243370e7-d511-4d9e-9893-32e6617a9512\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.785968 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttdgt\" (UniqueName: \"kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt\") pod \"243370e7-d511-4d9e-9893-32e6617a9512\" (UID: \"243370e7-d511-4d9e-9893-32e6617a9512\") " Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.787245 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities" (OuterVolumeSpecName: "utilities") pod "243370e7-d511-4d9e-9893-32e6617a9512" (UID: "243370e7-d511-4d9e-9893-32e6617a9512"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.793805 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt" (OuterVolumeSpecName: "kube-api-access-ttdgt") pod "243370e7-d511-4d9e-9893-32e6617a9512" (UID: "243370e7-d511-4d9e-9893-32e6617a9512"). InnerVolumeSpecName "kube-api-access-ttdgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.844396 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "243370e7-d511-4d9e-9893-32e6617a9512" (UID: "243370e7-d511-4d9e-9893-32e6617a9512"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.888835 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.888886 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttdgt\" (UniqueName: \"kubernetes.io/projected/243370e7-d511-4d9e-9893-32e6617a9512-kube-api-access-ttdgt\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:54 crc kubenswrapper[4972]: I0228 11:41:54.888902 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243370e7-d511-4d9e-9893-32e6617a9512-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.455711 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlq6d" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.455910 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlq6d" event={"ID":"243370e7-d511-4d9e-9893-32e6617a9512","Type":"ContainerDied","Data":"ed192b87f63482fdb288ea20a816ed5561088529e114841ce730188d9a974903"} Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.456592 4972 scope.go:117] "RemoveContainer" containerID="fa934fa9850e6749746907a61876adfd513d246fd13d65b7bae878cc40de4b86" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.486601 4972 scope.go:117] "RemoveContainer" containerID="35071658ab0e01b7e94b4c66b86bf35d103758c902ac157568aa846cdc97c0c6" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.511662 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.525562 4972 scope.go:117] "RemoveContainer" containerID="a93b099f73e82ec1fe1e6fdebebd8a90f6447d52e3cfe49398d181bd22c04cc0" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.530107 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mlq6d"] Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.806967 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" path="/var/lib/kubelet/pods/239a1f22-e163-4fa8-b9f0-0acc4d729fba/volumes" Feb 28 11:41:55 crc kubenswrapper[4972]: I0228 11:41:55.808246 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="243370e7-d511-4d9e-9893-32e6617a9512" path="/var/lib/kubelet/pods/243370e7-d511-4d9e-9893-32e6617a9512/volumes" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.165606 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537982-kgn58"] Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167368 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="extract-utilities" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167454 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="extract-utilities" Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167544 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="extract-content" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167559 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="extract-content" Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167587 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167600 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167624 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="extract-content" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167636 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="extract-content" Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167668 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167680 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: E0228 11:42:00.167701 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="extract-utilities" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.167714 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="extract-utilities" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.168140 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="239a1f22-e163-4fa8-b9f0-0acc4d729fba" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.168181 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="243370e7-d511-4d9e-9893-32e6617a9512" containerName="registry-server" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.169312 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.174353 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.174725 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.177942 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.178015 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537982-kgn58"] Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.334506 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9nlh\" (UniqueName: \"kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh\") pod \"auto-csr-approver-29537982-kgn58\" (UID: \"65b7ff79-c7ca-48bf-a36a-210158b98164\") " pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.436139 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9nlh\" (UniqueName: \"kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh\") pod \"auto-csr-approver-29537982-kgn58\" (UID: \"65b7ff79-c7ca-48bf-a36a-210158b98164\") " pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.469538 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9nlh\" (UniqueName: \"kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh\") pod \"auto-csr-approver-29537982-kgn58\" (UID: \"65b7ff79-c7ca-48bf-a36a-210158b98164\") " pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:00 crc kubenswrapper[4972]: I0228 11:42:00.532212 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:01 crc kubenswrapper[4972]: I0228 11:42:01.085499 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537982-kgn58"] Feb 28 11:42:01 crc kubenswrapper[4972]: I0228 11:42:01.553204 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537982-kgn58" event={"ID":"65b7ff79-c7ca-48bf-a36a-210158b98164","Type":"ContainerStarted","Data":"c899e3dec4434daa184c7fdb2be074d379ec07ea678e1c9edf9019127806981b"} Feb 28 11:42:02 crc kubenswrapper[4972]: I0228 11:42:02.565373 4972 generic.go:334] "Generic (PLEG): container finished" podID="65b7ff79-c7ca-48bf-a36a-210158b98164" containerID="0ca00e3e9aa68e5a7d16b43c9f6d72fc3c0d8b50975ad7e00c70aaf4c6e73707" exitCode=0 Feb 28 11:42:02 crc kubenswrapper[4972]: I0228 11:42:02.565585 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537982-kgn58" event={"ID":"65b7ff79-c7ca-48bf-a36a-210158b98164","Type":"ContainerDied","Data":"0ca00e3e9aa68e5a7d16b43c9f6d72fc3c0d8b50975ad7e00c70aaf4c6e73707"} Feb 28 11:42:02 crc kubenswrapper[4972]: I0228 11:42:02.790388 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:42:02 crc kubenswrapper[4972]: E0228 11:42:02.791290 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.132246 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.227542 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9nlh\" (UniqueName: \"kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh\") pod \"65b7ff79-c7ca-48bf-a36a-210158b98164\" (UID: \"65b7ff79-c7ca-48bf-a36a-210158b98164\") " Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.236136 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh" (OuterVolumeSpecName: "kube-api-access-r9nlh") pod "65b7ff79-c7ca-48bf-a36a-210158b98164" (UID: "65b7ff79-c7ca-48bf-a36a-210158b98164"). InnerVolumeSpecName "kube-api-access-r9nlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.330670 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9nlh\" (UniqueName: \"kubernetes.io/projected/65b7ff79-c7ca-48bf-a36a-210158b98164-kube-api-access-r9nlh\") on node \"crc\" DevicePath \"\"" Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.591311 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537982-kgn58" event={"ID":"65b7ff79-c7ca-48bf-a36a-210158b98164","Type":"ContainerDied","Data":"c899e3dec4434daa184c7fdb2be074d379ec07ea678e1c9edf9019127806981b"} Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.591364 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c899e3dec4434daa184c7fdb2be074d379ec07ea678e1c9edf9019127806981b" Feb 28 11:42:04 crc kubenswrapper[4972]: I0228 11:42:04.591428 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537982-kgn58" Feb 28 11:42:05 crc kubenswrapper[4972]: I0228 11:42:05.241455 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537976-6cw48"] Feb 28 11:42:05 crc kubenswrapper[4972]: I0228 11:42:05.254939 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537976-6cw48"] Feb 28 11:42:05 crc kubenswrapper[4972]: I0228 11:42:05.806731 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b619dc26-f8c5-4808-b31a-fd68eb4d1450" path="/var/lib/kubelet/pods/b619dc26-f8c5-4808-b31a-fd68eb4d1450/volumes" Feb 28 11:42:13 crc kubenswrapper[4972]: I0228 11:42:13.811984 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:42:13 crc kubenswrapper[4972]: E0228 11:42:13.818697 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:42:16 crc kubenswrapper[4972]: I0228 11:42:16.064157 4972 scope.go:117] "RemoveContainer" containerID="042679832b747a6e37be2d560ed26a4388ee522e1ef27008a7395e0fe15577a3" Feb 28 11:42:26 crc kubenswrapper[4972]: I0228 11:42:26.789391 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:42:26 crc kubenswrapper[4972]: E0228 11:42:26.790215 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:42:37 crc kubenswrapper[4972]: I0228 11:42:37.791038 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:42:37 crc kubenswrapper[4972]: E0228 11:42:37.794824 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:42:48 crc kubenswrapper[4972]: I0228 11:42:48.789423 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:42:48 crc kubenswrapper[4972]: E0228 11:42:48.790353 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.423495 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v89xz/must-gather-zb8lh"] Feb 28 11:42:52 crc kubenswrapper[4972]: E0228 11:42:52.424386 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b7ff79-c7ca-48bf-a36a-210158b98164" containerName="oc" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.424398 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b7ff79-c7ca-48bf-a36a-210158b98164" containerName="oc" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.424598 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b7ff79-c7ca-48bf-a36a-210158b98164" containerName="oc" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.425649 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.428367 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v89xz"/"kube-root-ca.crt" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.428615 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-v89xz"/"default-dockercfg-xd85v" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.428760 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v89xz"/"openshift-service-ca.crt" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.433010 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v89xz/must-gather-zb8lh"] Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.548398 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.548436 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d8th\" (UniqueName: \"kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.650704 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.650755 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d8th\" (UniqueName: \"kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.651152 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.681697 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d8th\" (UniqueName: \"kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th\") pod \"must-gather-zb8lh\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:52 crc kubenswrapper[4972]: I0228 11:42:52.743298 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:42:53 crc kubenswrapper[4972]: I0228 11:42:53.240984 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v89xz/must-gather-zb8lh"] Feb 28 11:42:54 crc kubenswrapper[4972]: I0228 11:42:54.168093 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/must-gather-zb8lh" event={"ID":"38760a3b-d6fe-466f-b10a-c2d3336082ae","Type":"ContainerStarted","Data":"9cc4879980ca1f8e9275270420781ebf6fade9b4c62c380ce26c1a2a2d5aabbd"} Feb 28 11:42:54 crc kubenswrapper[4972]: I0228 11:42:54.168625 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/must-gather-zb8lh" event={"ID":"38760a3b-d6fe-466f-b10a-c2d3336082ae","Type":"ContainerStarted","Data":"1987bcb7ecbe11c4d7e7920f5a3587354b67f7adcef516f405313a7edc82b515"} Feb 28 11:42:54 crc kubenswrapper[4972]: I0228 11:42:54.168658 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/must-gather-zb8lh" event={"ID":"38760a3b-d6fe-466f-b10a-c2d3336082ae","Type":"ContainerStarted","Data":"41288ee08f4bd61b46dd9c3e98d70099f26835a20199d9edc763bc407230bdae"} Feb 28 11:42:54 crc kubenswrapper[4972]: I0228 11:42:54.203489 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v89xz/must-gather-zb8lh" podStartSLOduration=2.203448605 podStartE2EDuration="2.203448605s" podCreationTimestamp="2026-02-28 11:42:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 11:42:54.188271718 +0000 UTC m=+4031.100279466" watchObservedRunningTime="2026-02-28 11:42:54.203448605 +0000 UTC m=+4031.115456353" Feb 28 11:42:56 crc kubenswrapper[4972]: I0228 11:42:56.916240 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v89xz/crc-debug-m6kb8"] Feb 28 11:42:56 crc kubenswrapper[4972]: I0228 11:42:56.918116 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.035131 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxg2\" (UniqueName: \"kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.035186 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.137350 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxg2\" (UniqueName: \"kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.137407 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.137611 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.159960 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxg2\" (UniqueName: \"kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2\") pod \"crc-debug-m6kb8\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: I0228 11:42:57.240313 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:42:57 crc kubenswrapper[4972]: W0228 11:42:57.270928 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86edef95_2e4d_48c2_a8c6_7bac0309bb21.slice/crio-5ada4d5beccedc48f44ed08a12c0d1248e2476f6f58a96c76aa4246c90b3abd7 WatchSource:0}: Error finding container 5ada4d5beccedc48f44ed08a12c0d1248e2476f6f58a96c76aa4246c90b3abd7: Status 404 returned error can't find the container with id 5ada4d5beccedc48f44ed08a12c0d1248e2476f6f58a96c76aa4246c90b3abd7 Feb 28 11:42:58 crc kubenswrapper[4972]: I0228 11:42:58.214784 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" event={"ID":"86edef95-2e4d-48c2-a8c6-7bac0309bb21","Type":"ContainerStarted","Data":"33dfba4f2daffd9b993abb3bb646656574b45aded0d0dd56b678c748c7ac833f"} Feb 28 11:42:58 crc kubenswrapper[4972]: I0228 11:42:58.215520 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" event={"ID":"86edef95-2e4d-48c2-a8c6-7bac0309bb21","Type":"ContainerStarted","Data":"5ada4d5beccedc48f44ed08a12c0d1248e2476f6f58a96c76aa4246c90b3abd7"} Feb 28 11:42:58 crc kubenswrapper[4972]: I0228 11:42:58.254428 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" podStartSLOduration=2.254401347 podStartE2EDuration="2.254401347s" podCreationTimestamp="2026-02-28 11:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-28 11:42:58.237027141 +0000 UTC m=+4035.149034919" watchObservedRunningTime="2026-02-28 11:42:58.254401347 +0000 UTC m=+4035.166409125" Feb 28 11:43:00 crc kubenswrapper[4972]: I0228 11:43:00.789190 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:43:00 crc kubenswrapper[4972]: E0228 11:43:00.790137 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:43:11 crc kubenswrapper[4972]: I0228 11:43:11.790444 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:43:11 crc kubenswrapper[4972]: E0228 11:43:11.792862 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:43:25 crc kubenswrapper[4972]: I0228 11:43:25.789519 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:43:25 crc kubenswrapper[4972]: E0228 11:43:25.790278 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:43:29 crc kubenswrapper[4972]: I0228 11:43:29.603855 4972 generic.go:334] "Generic (PLEG): container finished" podID="86edef95-2e4d-48c2-a8c6-7bac0309bb21" containerID="33dfba4f2daffd9b993abb3bb646656574b45aded0d0dd56b678c748c7ac833f" exitCode=0 Feb 28 11:43:29 crc kubenswrapper[4972]: I0228 11:43:29.603940 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" event={"ID":"86edef95-2e4d-48c2-a8c6-7bac0309bb21","Type":"ContainerDied","Data":"33dfba4f2daffd9b993abb3bb646656574b45aded0d0dd56b678c748c7ac833f"} Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.700777 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.733825 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-m6kb8"] Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.741345 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-m6kb8"] Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.866546 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kxg2\" (UniqueName: \"kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2\") pod \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.867065 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host\") pod \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\" (UID: \"86edef95-2e4d-48c2-a8c6-7bac0309bb21\") " Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.867134 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host" (OuterVolumeSpecName: "host") pod "86edef95-2e4d-48c2-a8c6-7bac0309bb21" (UID: "86edef95-2e4d-48c2-a8c6-7bac0309bb21"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.867736 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86edef95-2e4d-48c2-a8c6-7bac0309bb21-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.872606 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2" (OuterVolumeSpecName: "kube-api-access-4kxg2") pod "86edef95-2e4d-48c2-a8c6-7bac0309bb21" (UID: "86edef95-2e4d-48c2-a8c6-7bac0309bb21"). InnerVolumeSpecName "kube-api-access-4kxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:43:30 crc kubenswrapper[4972]: I0228 11:43:30.969234 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kxg2\" (UniqueName: \"kubernetes.io/projected/86edef95-2e4d-48c2-a8c6-7bac0309bb21-kube-api-access-4kxg2\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.624314 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ada4d5beccedc48f44ed08a12c0d1248e2476f6f58a96c76aa4246c90b3abd7" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.624389 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-m6kb8" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.799163 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86edef95-2e4d-48c2-a8c6-7bac0309bb21" path="/var/lib/kubelet/pods/86edef95-2e4d-48c2-a8c6-7bac0309bb21/volumes" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.929677 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v89xz/crc-debug-lnn47"] Feb 28 11:43:31 crc kubenswrapper[4972]: E0228 11:43:31.930062 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86edef95-2e4d-48c2-a8c6-7bac0309bb21" containerName="container-00" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.930077 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="86edef95-2e4d-48c2-a8c6-7bac0309bb21" containerName="container-00" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.930251 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="86edef95-2e4d-48c2-a8c6-7bac0309bb21" containerName="container-00" Feb 28 11:43:31 crc kubenswrapper[4972]: I0228 11:43:31.934650 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.088100 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7bdm\" (UniqueName: \"kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.088239 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.189935 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.190045 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.190050 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7bdm\" (UniqueName: \"kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.215107 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7bdm\" (UniqueName: \"kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm\") pod \"crc-debug-lnn47\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.251043 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.635103 4972 generic.go:334] "Generic (PLEG): container finished" podID="10219974-d7ff-4eb6-aae5-158a274dd645" containerID="a1559d67ed9e2a1777dff6498023985e601e33568c16d5ecfa6aaa8b94ec6b06" exitCode=0 Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.635200 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-lnn47" event={"ID":"10219974-d7ff-4eb6-aae5-158a274dd645","Type":"ContainerDied","Data":"a1559d67ed9e2a1777dff6498023985e601e33568c16d5ecfa6aaa8b94ec6b06"} Feb 28 11:43:32 crc kubenswrapper[4972]: I0228 11:43:32.635506 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-lnn47" event={"ID":"10219974-d7ff-4eb6-aae5-158a274dd645","Type":"ContainerStarted","Data":"b439f5f32892df0f5bfbbac2e686c28ca695dc3386df991824691d78398dea8c"} Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.092325 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-lnn47"] Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.101264 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-lnn47"] Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.749350 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.819706 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host\") pod \"10219974-d7ff-4eb6-aae5-158a274dd645\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.819771 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7bdm\" (UniqueName: \"kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm\") pod \"10219974-d7ff-4eb6-aae5-158a274dd645\" (UID: \"10219974-d7ff-4eb6-aae5-158a274dd645\") " Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.819878 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host" (OuterVolumeSpecName: "host") pod "10219974-d7ff-4eb6-aae5-158a274dd645" (UID: "10219974-d7ff-4eb6-aae5-158a274dd645"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.820181 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10219974-d7ff-4eb6-aae5-158a274dd645-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.830705 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm" (OuterVolumeSpecName: "kube-api-access-h7bdm") pod "10219974-d7ff-4eb6-aae5-158a274dd645" (UID: "10219974-d7ff-4eb6-aae5-158a274dd645"). InnerVolumeSpecName "kube-api-access-h7bdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:43:33 crc kubenswrapper[4972]: I0228 11:43:33.921766 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7bdm\" (UniqueName: \"kubernetes.io/projected/10219974-d7ff-4eb6-aae5-158a274dd645-kube-api-access-h7bdm\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.339336 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v89xz/crc-debug-62lz6"] Feb 28 11:43:34 crc kubenswrapper[4972]: E0228 11:43:34.339872 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10219974-d7ff-4eb6-aae5-158a274dd645" containerName="container-00" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.339893 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="10219974-d7ff-4eb6-aae5-158a274dd645" containerName="container-00" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.340138 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="10219974-d7ff-4eb6-aae5-158a274dd645" containerName="container-00" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.341668 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.433125 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.433206 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n4qx\" (UniqueName: \"kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.535290 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.535371 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n4qx\" (UniqueName: \"kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.535490 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.552384 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n4qx\" (UniqueName: \"kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx\") pod \"crc-debug-62lz6\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.657967 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.658753 4972 scope.go:117] "RemoveContainer" containerID="a1559d67ed9e2a1777dff6498023985e601e33568c16d5ecfa6aaa8b94ec6b06" Feb 28 11:43:34 crc kubenswrapper[4972]: I0228 11:43:34.658856 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-lnn47" Feb 28 11:43:34 crc kubenswrapper[4972]: W0228 11:43:34.698485 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f940226_767c_4600_9475_406f62eb401b.slice/crio-a5d2b6f9087ca2de4ac5189932b0238f1ac406108df355f316725a3a74fc0106 WatchSource:0}: Error finding container a5d2b6f9087ca2de4ac5189932b0238f1ac406108df355f316725a3a74fc0106: Status 404 returned error can't find the container with id a5d2b6f9087ca2de4ac5189932b0238f1ac406108df355f316725a3a74fc0106 Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.670778 4972 generic.go:334] "Generic (PLEG): container finished" podID="5f940226-767c-4600-9475-406f62eb401b" containerID="604d57178f277cd5c2c0cf8e76df7c3c6935cde0326a00aa6ee9f93bbf38d70f" exitCode=0 Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.670883 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-62lz6" event={"ID":"5f940226-767c-4600-9475-406f62eb401b","Type":"ContainerDied","Data":"604d57178f277cd5c2c0cf8e76df7c3c6935cde0326a00aa6ee9f93bbf38d70f"} Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.671301 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/crc-debug-62lz6" event={"ID":"5f940226-767c-4600-9475-406f62eb401b","Type":"ContainerStarted","Data":"a5d2b6f9087ca2de4ac5189932b0238f1ac406108df355f316725a3a74fc0106"} Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.715926 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-62lz6"] Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.726164 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v89xz/crc-debug-62lz6"] Feb 28 11:43:35 crc kubenswrapper[4972]: I0228 11:43:35.804615 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10219974-d7ff-4eb6-aae5-158a274dd645" path="/var/lib/kubelet/pods/10219974-d7ff-4eb6-aae5-158a274dd645/volumes" Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.790564 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.881747 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n4qx\" (UniqueName: \"kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx\") pod \"5f940226-767c-4600-9475-406f62eb401b\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.881867 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host\") pod \"5f940226-767c-4600-9475-406f62eb401b\" (UID: \"5f940226-767c-4600-9475-406f62eb401b\") " Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.882016 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host" (OuterVolumeSpecName: "host") pod "5f940226-767c-4600-9475-406f62eb401b" (UID: "5f940226-767c-4600-9475-406f62eb401b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.882497 4972 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f940226-767c-4600-9475-406f62eb401b-host\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.887316 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx" (OuterVolumeSpecName: "kube-api-access-9n4qx") pod "5f940226-767c-4600-9475-406f62eb401b" (UID: "5f940226-767c-4600-9475-406f62eb401b"). InnerVolumeSpecName "kube-api-access-9n4qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:43:36 crc kubenswrapper[4972]: I0228 11:43:36.985288 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n4qx\" (UniqueName: \"kubernetes.io/projected/5f940226-767c-4600-9475-406f62eb401b-kube-api-access-9n4qx\") on node \"crc\" DevicePath \"\"" Feb 28 11:43:37 crc kubenswrapper[4972]: I0228 11:43:37.691684 4972 scope.go:117] "RemoveContainer" containerID="604d57178f277cd5c2c0cf8e76df7c3c6935cde0326a00aa6ee9f93bbf38d70f" Feb 28 11:43:37 crc kubenswrapper[4972]: I0228 11:43:37.691745 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/crc-debug-62lz6" Feb 28 11:43:37 crc kubenswrapper[4972]: I0228 11:43:37.789316 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:43:37 crc kubenswrapper[4972]: E0228 11:43:37.789603 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:43:37 crc kubenswrapper[4972]: I0228 11:43:37.798957 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f940226-767c-4600-9475-406f62eb401b" path="/var/lib/kubelet/pods/5f940226-767c-4600-9475-406f62eb401b/volumes" Feb 28 11:43:51 crc kubenswrapper[4972]: I0228 11:43:51.790319 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:43:51 crc kubenswrapper[4972]: E0228 11:43:51.791756 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.155982 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537984-gpnsb"] Feb 28 11:44:00 crc kubenswrapper[4972]: E0228 11:44:00.158754 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f940226-767c-4600-9475-406f62eb401b" containerName="container-00" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.158937 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f940226-767c-4600-9475-406f62eb401b" containerName="container-00" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.159536 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f940226-767c-4600-9475-406f62eb401b" containerName="container-00" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.161118 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.166233 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.166605 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.166843 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.168873 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537984-gpnsb"] Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.245166 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhkq2\" (UniqueName: \"kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2\") pod \"auto-csr-approver-29537984-gpnsb\" (UID: \"ef06ae10-f323-4fb0-b44c-0a8f549e4915\") " pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.346868 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhkq2\" (UniqueName: \"kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2\") pod \"auto-csr-approver-29537984-gpnsb\" (UID: \"ef06ae10-f323-4fb0-b44c-0a8f549e4915\") " pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.370905 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhkq2\" (UniqueName: \"kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2\") pod \"auto-csr-approver-29537984-gpnsb\" (UID: \"ef06ae10-f323-4fb0-b44c-0a8f549e4915\") " pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:00 crc kubenswrapper[4972]: I0228 11:44:00.516246 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:01 crc kubenswrapper[4972]: I0228 11:44:01.036687 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537984-gpnsb"] Feb 28 11:44:01 crc kubenswrapper[4972]: I0228 11:44:01.037665 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:44:01 crc kubenswrapper[4972]: I0228 11:44:01.957488 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" event={"ID":"ef06ae10-f323-4fb0-b44c-0a8f549e4915","Type":"ContainerStarted","Data":"3ced010d17238ed54f0e3e12af42f75c9e83361656bdb285c03de90f8f33a34d"} Feb 28 11:44:02 crc kubenswrapper[4972]: I0228 11:44:02.966893 4972 generic.go:334] "Generic (PLEG): container finished" podID="ef06ae10-f323-4fb0-b44c-0a8f549e4915" containerID="d2e67657f393e73b1281dcacd11e5a9f4e97c47e6bf3bc6b1ec3bcb6c0b39a4f" exitCode=0 Feb 28 11:44:02 crc kubenswrapper[4972]: I0228 11:44:02.967092 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" event={"ID":"ef06ae10-f323-4fb0-b44c-0a8f549e4915","Type":"ContainerDied","Data":"d2e67657f393e73b1281dcacd11e5a9f4e97c47e6bf3bc6b1ec3bcb6c0b39a4f"} Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.306809 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.337801 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhkq2\" (UniqueName: \"kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2\") pod \"ef06ae10-f323-4fb0-b44c-0a8f549e4915\" (UID: \"ef06ae10-f323-4fb0-b44c-0a8f549e4915\") " Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.347634 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2" (OuterVolumeSpecName: "kube-api-access-hhkq2") pod "ef06ae10-f323-4fb0-b44c-0a8f549e4915" (UID: "ef06ae10-f323-4fb0-b44c-0a8f549e4915"). InnerVolumeSpecName "kube-api-access-hhkq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.440775 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhkq2\" (UniqueName: \"kubernetes.io/projected/ef06ae10-f323-4fb0-b44c-0a8f549e4915-kube-api-access-hhkq2\") on node \"crc\" DevicePath \"\"" Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.986123 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" event={"ID":"ef06ae10-f323-4fb0-b44c-0a8f549e4915","Type":"ContainerDied","Data":"3ced010d17238ed54f0e3e12af42f75c9e83361656bdb285c03de90f8f33a34d"} Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.986182 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ced010d17238ed54f0e3e12af42f75c9e83361656bdb285c03de90f8f33a34d" Feb 28 11:44:04 crc kubenswrapper[4972]: I0228 11:44:04.986209 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537984-gpnsb" Feb 28 11:44:05 crc kubenswrapper[4972]: I0228 11:44:05.380625 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537978-42wm9"] Feb 28 11:44:05 crc kubenswrapper[4972]: I0228 11:44:05.392327 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537978-42wm9"] Feb 28 11:44:05 crc kubenswrapper[4972]: I0228 11:44:05.789699 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:44:05 crc kubenswrapper[4972]: E0228 11:44:05.790101 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:05 crc kubenswrapper[4972]: I0228 11:44:05.803492 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4142236e-468b-4859-b322-4899a912363f" path="/var/lib/kubelet/pods/4142236e-468b-4859-b322-4899a912363f/volumes" Feb 28 11:44:16 crc kubenswrapper[4972]: I0228 11:44:16.236723 4972 scope.go:117] "RemoveContainer" containerID="78d2c2e3d0cc3e115c09ffd2b8416df25ea52397c018e2032d7b0e66ac28a300" Feb 28 11:44:16 crc kubenswrapper[4972]: I0228 11:44:16.788882 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:44:16 crc kubenswrapper[4972]: E0228 11:44:16.789324 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.062692 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f94fccfd-62888_e50121ef-8724-4a1d-8ee2-badc2c40df58/barbican-api/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.122032 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f94fccfd-62888_e50121ef-8724-4a1d-8ee2-badc2c40df58/barbican-api-log/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.257474 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64d46dc98-5dqwn_fed4da99-54b6-4444-8f0f-4489a5286e4d/barbican-keystone-listener/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.269326 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-64d46dc98-5dqwn_fed4da99-54b6-4444-8f0f-4489a5286e4d/barbican-keystone-listener-log/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.367427 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9686d86c-lmjsc_31309c10-9f68-4aed-ab99-dcef67e12b92/barbican-worker/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.432650 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9686d86c-lmjsc_31309c10-9f68-4aed-ab99-dcef67e12b92/barbican-worker-log/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.535212 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fl5fc_e9d8a507-2de3-4624-ad38-5d50c0c50874/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.637825 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/ceilometer-central-agent/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.706232 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/ceilometer-notification-agent/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.729032 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/proxy-httpd/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.799412 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a6e6077-551d-4ef1-b669-3c5bac893957/sg-core/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.938258 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d0cdd4b-c98a-4297-b44d-9fcb336aecf0/cinder-api-log/0.log" Feb 28 11:44:17 crc kubenswrapper[4972]: I0228 11:44:17.947294 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d0cdd4b-c98a-4297-b44d-9fcb336aecf0/cinder-api/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.097746 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_84de24a3-0eae-4ec5-af31-a515e98ef012/cinder-scheduler/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.129816 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_84de24a3-0eae-4ec5-af31-a515e98ef012/probe/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.252236 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zxdwz_51ff636b-adde-4a49-a96d-bf1275e5b200/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.754392 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/init/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.780450 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-q2clh_5dc25a96-f5f2-491d-b342-41bb674572ba/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.913883 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/init/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.990837 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-2qbql_977a02d6-198e-4d62-a2aa-9f83ae367dc6/dnsmasq-dns/0.log" Feb 28 11:44:18 crc kubenswrapper[4972]: I0228 11:44:18.991563 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-wfqdt_c2bf8774-78c6-4d38-9d1b-6f24cf1d95c3/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.139007 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e00dd1be-06c1-4c4a-979e-3cb562e7741e/glance-httpd/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.184238 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e00dd1be-06c1-4c4a-979e-3cb562e7741e/glance-log/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.353311 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_35925034-cd2f-4537-8225-0ab87fd3a47e/glance-log/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.362210 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_35925034-cd2f-4537-8225-0ab87fd3a47e/glance-httpd/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.506374 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c6cf8f456-722rs_a24fd1fd-763d-4106-87b0-805e2f1fc6c5/horizon/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.573365 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-grxc8_585066f8-afbf-42ac-8163-895ef6b76ec5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.832657 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9kfmp_8a963845-f99d-488b-81c8-b13e44cc5ff2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:19 crc kubenswrapper[4972]: I0228 11:44:19.866719 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c6cf8f456-722rs_a24fd1fd-763d-4106-87b0-805e2f1fc6c5/horizon-log/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.065595 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29537941-8tjbq_eaa3c5ff-5646-444a-b502-6052100adb2e/keystone-cron/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.086025 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-66b556fdbd-nljfl_cdfe3155-40f6-48df-accc-e7667625cd12/keystone-api/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.280320 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-b6wzz_1a18fc65-40ad-474f-b28d-76e3d12cb6df/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.295550 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_15239450-dea0-4aab-8fe7-1d891d57afab/kube-state-metrics/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.562022 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69dd77478c-x64n9_b7d2ee5d-2b90-4c97-a931-45992f1f50d1/neutron-api/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.679884 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-69dd77478c-x64n9_b7d2ee5d-2b90-4c97-a931-45992f1f50d1/neutron-httpd/0.log" Feb 28 11:44:20 crc kubenswrapper[4972]: I0228 11:44:20.830268 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8hx2g_ae3ba487-6066-46c6-b9f2-5acc9810d3bf/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.275206 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_64af5818-baad-43d6-abb6-644fb18b3b5e/nova-api-log/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.329112 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3e68715e-2416-4459-8c91-6368dbedb67a/nova-cell0-conductor-conductor/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.685744 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_3563bc2b-2c86-4824-b238-e8a60f45c60e/nova-cell1-conductor-conductor/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.698734 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_70d1c7cf-2d6e-4076-b7cc-8752e5172008/nova-cell1-novncproxy-novncproxy/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.723163 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_64af5818-baad-43d6-abb6-644fb18b3b5e/nova-api-api/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.938014 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dqvpd_d0655084-fd3f-4088-bb4e-755a5db445fe/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:21 crc kubenswrapper[4972]: I0228 11:44:21.968410 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_92b274e8-6674-42a1-81f3-302d28e22926/nova-metadata-log/0.log" Feb 28 11:44:22 crc kubenswrapper[4972]: I0228 11:44:22.389235 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/mysql-bootstrap/0.log" Feb 28 11:44:22 crc kubenswrapper[4972]: I0228 11:44:22.447716 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_44f084cb-b672-497c-8c66-23785c9971f7/nova-scheduler-scheduler/0.log" Feb 28 11:44:22 crc kubenswrapper[4972]: I0228 11:44:22.564127 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/mysql-bootstrap/0.log" Feb 28 11:44:22 crc kubenswrapper[4972]: I0228 11:44:22.569524 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_44e7c813-beaa-4015-a581-afcaea85077c/galera/0.log" Feb 28 11:44:22 crc kubenswrapper[4972]: I0228 11:44:22.777797 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/mysql-bootstrap/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.008901 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/mysql-bootstrap/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.021287 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_57d374b4-97c1-4b80-8f4b-95338fe4f385/galera/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.181574 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f069c4ff-a971-4849-9d09-4ae37dca70fc/openstackclient/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.252670 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jl8t9_0be6aa95-e5d9-4e85-9991-d58945420d69/openstack-network-exporter/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.276447 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_92b274e8-6674-42a1-81f3-302d28e22926/nova-metadata-metadata/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.424913 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server-init/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.598825 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovs-vswitchd/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.679978 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.680850 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5wgfj_7eb839e4-8819-45ba-9afd-312cc02a6964/ovsdb-server-init/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.835634 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p6jmv_d90126a7-5923-45bd-9e61-29d1e2e3f469/ovn-controller/0.log" Feb 28 11:44:23 crc kubenswrapper[4972]: I0228 11:44:23.946993 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-k9z8j_d2cb5f2e-f9b1-41fb-8a63-03bebe97019b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.045123 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8ccd0a65-b9b9-4b45-ba11-97903d81f938/ovn-northd/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.055799 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8ccd0a65-b9b9-4b45-ba11-97903d81f938/openstack-network-exporter/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.125024 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6ba1065c-4ef3-473c-9576-d673f19a75a1/openstack-network-exporter/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.243630 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6ba1065c-4ef3-473c-9576-d673f19a75a1/ovsdbserver-nb/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.322072 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec76b1b5-ea80-49cb-ae7c-0190a52640a1/openstack-network-exporter/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.355198 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ec76b1b5-ea80-49cb-ae7c-0190a52640a1/ovsdbserver-sb/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.584979 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bb67f8988-tt9dr_bae387b2-3007-4c94-9a6f-255bd137d7c0/placement-log/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.593984 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bb67f8988-tt9dr_bae387b2-3007-4c94-9a6f-255bd137d7c0/placement-api/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.695740 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/setup-container/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.894478 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/setup-container/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.971732 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/setup-container/0.log" Feb 28 11:44:24 crc kubenswrapper[4972]: I0228 11:44:24.976006 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66540e7-d517-4b5d-bd92-efdd423596fb/rabbitmq/0.log" Feb 28 11:44:25 crc kubenswrapper[4972]: I0228 11:44:25.106561 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/setup-container/0.log" Feb 28 11:44:25 crc kubenswrapper[4972]: I0228 11:44:25.176817 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xdp6w_dea3597c-d878-41dd-8952-dda5b8db255e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:25 crc kubenswrapper[4972]: I0228 11:44:25.177734 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6a821361-b8f7-4511-9f07-9f05f1b8425d/rabbitmq/0.log" Feb 28 11:44:25 crc kubenswrapper[4972]: I0228 11:44:25.503708 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-bvts2_a6c60946-e11b-4644-8742-fb1225f8d690/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:25 crc kubenswrapper[4972]: I0228 11:44:25.569439 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jhw6f_a87a4176-9953-45b9-8db1-824b39dd2e4d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:26 crc kubenswrapper[4972]: I0228 11:44:26.710220 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dskz4_a19c5936-e88b-4336-814d-8b0053fe2247/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:26 crc kubenswrapper[4972]: I0228 11:44:26.761086 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nqj8m_41b5f77f-8955-4d82-8d40-08c398090bc7/ssh-known-hosts-edpm-deployment/0.log" Feb 28 11:44:26 crc kubenswrapper[4972]: I0228 11:44:26.947984 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98999f697-8tcfb_09caa13b-95cf-42c1-a35d-7c053e069909/proxy-server/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.015303 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98999f697-8tcfb_09caa13b-95cf-42c1-a35d-7c053e069909/proxy-httpd/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.091739 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2f84x_28171537-6c72-4fe2-9e23-709a2f0824ae/swift-ring-rebalance/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.140769 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-auditor/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.205402 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-reaper/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.348319 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-replicator/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.419501 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-auditor/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.444490 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/account-server/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.464660 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-replicator/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.561539 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-server/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.649880 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/container-updater/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.658192 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-auditor/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.683646 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-expirer/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.770265 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-replicator/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.883283 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-updater/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.903601 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/object-server/0.log" Feb 28 11:44:27 crc kubenswrapper[4972]: I0228 11:44:27.903635 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/rsync/0.log" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.478796 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4376e6f8-3008-4aeb-b049-54ee633833f2/swift-recon-cron/0.log" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.492129 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-shq6f_a667491e-8728-4640-86b9-410f2cf770c9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.703874 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fb460ac4-a796-4700-ab43-565ca0308a20/test-operator-logs-container/0.log" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.705890 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_50684e40-2fa2-4cb3-a176-374ab716698c/tempest-tests-tempest-tests-runner/0.log" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.788971 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:44:28 crc kubenswrapper[4972]: E0228 11:44:28.789267 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:28 crc kubenswrapper[4972]: I0228 11:44:28.899621 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4rrfz_6436504e-d8a5-4dc9-b500-810a55ef3bbb/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 28 11:44:38 crc kubenswrapper[4972]: I0228 11:44:38.203771 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_56e07ece-d896-4107-b9b1-6f24b64f7985/memcached/0.log" Feb 28 11:44:39 crc kubenswrapper[4972]: I0228 11:44:39.789420 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:44:39 crc kubenswrapper[4972]: E0228 11:44:39.789970 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:52 crc kubenswrapper[4972]: I0228 11:44:52.789553 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:44:52 crc kubenswrapper[4972]: E0228 11:44:52.790415 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:44:57 crc kubenswrapper[4972]: I0228 11:44:57.790197 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.426094 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.447811 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.536429 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.717516 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/pull/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.722168 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/extract/0.log" Feb 28 11:44:58 crc kubenswrapper[4972]: I0228 11:44:58.759172 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_38a344b2f93d4de9048cd71831f2eff772c73314b8e22231a68493657bm8kph_4c8c46d3-deb0-4552-9e73-d565dd09da2a/util/0.log" Feb 28 11:44:59 crc kubenswrapper[4972]: I0228 11:44:59.195973 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-55cc45767f-krrsg_61ba5500-d464-418c-9248-bc05118b0632/manager/0.log" Feb 28 11:44:59 crc kubenswrapper[4972]: I0228 11:44:59.524201 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5fb5494549-ft8w9_fe59fc2c-7e44-4b09-9d79-415c90f1d3fa/manager/0.log" Feb 28 11:44:59 crc kubenswrapper[4972]: I0228 11:44:59.707057 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-585b788787-j2pnv_3abb4675-640d-4b5c-b3fb-55279630aecd/manager/0.log" Feb 28 11:44:59 crc kubenswrapper[4972]: I0228 11:44:59.945342 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7db95d7ffb-wdlpp_fb530b5d-27e7-4973-9fb5-41b935e9c0a0/manager/0.log" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.149251 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b"] Feb 28 11:45:00 crc kubenswrapper[4972]: E0228 11:45:00.149657 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef06ae10-f323-4fb0-b44c-0a8f549e4915" containerName="oc" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.149671 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef06ae10-f323-4fb0-b44c-0a8f549e4915" containerName="oc" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.149847 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef06ae10-f323-4fb0-b44c-0a8f549e4915" containerName="oc" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.150412 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.154615 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.154619 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.159944 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b"] Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.284751 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.284805 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw9s4\" (UniqueName: \"kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.284863 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.335904 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-8784b4656-hhmnw_20457d28-02f6-4d31-9938-dbdec910d3c8/manager/0.log" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.386018 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.386075 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw9s4\" (UniqueName: \"kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.386144 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.387301 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.400178 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.402932 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw9s4\" (UniqueName: \"kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4\") pod \"collect-profiles-29537985-jgc6b\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.471650 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.657861 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-78b64779b9-xt2pl_15d3191f-413a-4806-8e55-9dd19db4c335/manager/0.log" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.727707 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-c77466965-bkt68_0c55bc45-0607-477d-893f-5782b00cffb2/manager/0.log" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.848533 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6f6f57b9b6-rxzp8_af73edbb-e925-4e12-9cce-58fe2292a497/manager/0.log" Feb 28 11:45:00 crc kubenswrapper[4972]: I0228 11:45:00.947433 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b"] Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.059776 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-745fc45789-25bfz_f2d950d1-acba-46f8-99c7-a932569c1c49/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.325709 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-768f998cf4-4mxwn_e099f5ae-f38b-4cb6-8be1-bbd8f91c7e86/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.513041 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-cc79fdffd-lgrm6_bae14f0d-c03b-4bf4-9812-2a4cdcae1d94/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.514580 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c67ff7674-v8vzt_adaec4ab-a8f3-4719-809d-b102d8e14b7b/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.571303 4972 generic.go:334] "Generic (PLEG): container finished" podID="bdb996af-96ee-43e3-9ef8-c7e124611f54" containerID="0a06b8043f7722c31284c291b3b8cd5f52c3c61182c4dd2d46aa0ea9612d526c" exitCode=0 Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.571397 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" event={"ID":"bdb996af-96ee-43e3-9ef8-c7e124611f54","Type":"ContainerDied","Data":"0a06b8043f7722c31284c291b3b8cd5f52c3c61182c4dd2d46aa0ea9612d526c"} Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.571542 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" event={"ID":"bdb996af-96ee-43e3-9ef8-c7e124611f54","Type":"ContainerStarted","Data":"649ffefac27dae1bccd25ac0735d5c3b958b10bbec9ae760b92304a17d20c60f"} Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.620056 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-768c8b45bb-w987c_7885ab2e-6366-4513-8904-a982144ac4c4/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.723894 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-c5677dc5d-6t4gw_473197b6-6493-4509-920b-ab1be61070f9/manager/0.log" Feb 28 11:45:01 crc kubenswrapper[4972]: I0228 11:45:01.993064 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6cdcc94fb7-qfx4x_cf37191c-db05-439a-9353-1e15f8147289/operator/0.log" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.181886 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pvw8x_8e82abb6-3796-42c1-bfe7-688f9117943f/registry-server/0.log" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.429371 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-684c7d77b-x7wnv_7068aaa9-0410-48a7-9122-53dfad3b2e67/manager/0.log" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.557774 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-bff955cc4-gxsgb_9c3297f2-e6e9-41e0-b299-fae7573d16cc/manager/0.log" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.726505 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-8ksfr_5c4dd163-6c95-4f91-8c40-b242a0d191e3/operator/0.log" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.926897 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:02 crc kubenswrapper[4972]: I0228 11:45:02.929797 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-55f4bf89cb-jm2wm_510043b9-9bba-49a8-9902-e37f4564bbe8/manager/0.log" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.035331 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume\") pod \"bdb996af-96ee-43e3-9ef8-c7e124611f54\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.035440 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume\") pod \"bdb996af-96ee-43e3-9ef8-c7e124611f54\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.035637 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw9s4\" (UniqueName: \"kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4\") pod \"bdb996af-96ee-43e3-9ef8-c7e124611f54\" (UID: \"bdb996af-96ee-43e3-9ef8-c7e124611f54\") " Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.036334 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume" (OuterVolumeSpecName: "config-volume") pod "bdb996af-96ee-43e3-9ef8-c7e124611f54" (UID: "bdb996af-96ee-43e3-9ef8-c7e124611f54"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.041291 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bdb996af-96ee-43e3-9ef8-c7e124611f54" (UID: "bdb996af-96ee-43e3-9ef8-c7e124611f54"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.041864 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4" (OuterVolumeSpecName: "kube-api-access-bw9s4") pod "bdb996af-96ee-43e3-9ef8-c7e124611f54" (UID: "bdb996af-96ee-43e3-9ef8-c7e124611f54"). InnerVolumeSpecName "kube-api-access-bw9s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.139670 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw9s4\" (UniqueName: \"kubernetes.io/projected/bdb996af-96ee-43e3-9ef8-c7e124611f54-kube-api-access-bw9s4\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.139713 4972 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bdb996af-96ee-43e3-9ef8-c7e124611f54-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.139725 4972 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bdb996af-96ee-43e3-9ef8-c7e124611f54-config-volume\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.191378 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8467ccb4c8-d9mct_81c08d80-fa23-44c5-9ddf-65dc85b18696/manager/0.log" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.220404 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-56dc67d744-trxff_2da6f8cf-2f39-4614-be52-b6598a919217/manager/0.log" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.473498 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-65c9f4f6b-vpxcf_65307cfe-55b4-486c-b154-2196bd0a1012/manager/0.log" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.588177 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" event={"ID":"bdb996af-96ee-43e3-9ef8-c7e124611f54","Type":"ContainerDied","Data":"649ffefac27dae1bccd25ac0735d5c3b958b10bbec9ae760b92304a17d20c60f"} Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.588499 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649ffefac27dae1bccd25ac0735d5c3b958b10bbec9ae760b92304a17d20c60f" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.588561 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29537985-jgc6b" Feb 28 11:45:03 crc kubenswrapper[4972]: I0228 11:45:03.619003 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cf56d77c9-f7stz_8cce4f5d-78aa-44c4-8c60-8bce681ada4e/manager/0.log" Feb 28 11:45:04 crc kubenswrapper[4972]: I0228 11:45:04.008549 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz"] Feb 28 11:45:04 crc kubenswrapper[4972]: I0228 11:45:04.017035 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29537940-6xkzz"] Feb 28 11:45:05 crc kubenswrapper[4972]: I0228 11:45:05.788922 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:45:05 crc kubenswrapper[4972]: E0228 11:45:05.789398 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:45:05 crc kubenswrapper[4972]: I0228 11:45:05.798908 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8edbd92b-6e5b-4c0f-b5d0-517a31b06960" path="/var/lib/kubelet/pods/8edbd92b-6e5b-4c0f-b5d0-517a31b06960/volumes" Feb 28 11:45:08 crc kubenswrapper[4972]: I0228 11:45:08.914360 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6fb74c6d59-x6hp2_9c739be1-15cd-4044-8cce-c3d2750d030e/manager/0.log" Feb 28 11:45:16 crc kubenswrapper[4972]: I0228 11:45:16.381083 4972 scope.go:117] "RemoveContainer" containerID="ae89a2b6508bbc454200bbf4bfcdce2c7398ae70184ef633f01be25849c96dee" Feb 28 11:45:16 crc kubenswrapper[4972]: I0228 11:45:16.789296 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:45:16 crc kubenswrapper[4972]: E0228 11:45:16.789821 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:45:27 crc kubenswrapper[4972]: I0228 11:45:27.803674 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gzmqk_8066e21c-fa0c-497d-8bc8-a27d07dab5b4/control-plane-machine-set-operator/0.log" Feb 28 11:45:27 crc kubenswrapper[4972]: I0228 11:45:27.960721 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xwmd_c9173e02-eddc-40e3-86ab-0f0e8a8f50ac/kube-rbac-proxy/0.log" Feb 28 11:45:28 crc kubenswrapper[4972]: I0228 11:45:28.009673 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xwmd_c9173e02-eddc-40e3-86ab-0f0e8a8f50ac/machine-api-operator/0.log" Feb 28 11:45:29 crc kubenswrapper[4972]: I0228 11:45:29.789176 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:45:29 crc kubenswrapper[4972]: E0228 11:45:29.789920 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.774884 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:38 crc kubenswrapper[4972]: E0228 11:45:38.775782 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb996af-96ee-43e3-9ef8-c7e124611f54" containerName="collect-profiles" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.775795 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb996af-96ee-43e3-9ef8-c7e124611f54" containerName="collect-profiles" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.776003 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb996af-96ee-43e3-9ef8-c7e124611f54" containerName="collect-profiles" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.777312 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.784896 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.921476 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.921565 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:38 crc kubenswrapper[4972]: I0228 11:45:38.921643 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq2jn\" (UniqueName: \"kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.024654 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.024744 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.024801 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq2jn\" (UniqueName: \"kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.025693 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.025737 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.045058 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq2jn\" (UniqueName: \"kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn\") pod \"redhat-marketplace-bx5n6\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.096675 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.580799 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.958596 4972 generic.go:334] "Generic (PLEG): container finished" podID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerID="5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398" exitCode=0 Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.958678 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerDied","Data":"5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398"} Feb 28 11:45:39 crc kubenswrapper[4972]: I0228 11:45:39.958724 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerStarted","Data":"af424507ce75d4adbec1446182c3f99a59f6b51eb222dd333846951b6661ca56"} Feb 28 11:45:40 crc kubenswrapper[4972]: I0228 11:45:40.789190 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:45:40 crc kubenswrapper[4972]: E0228 11:45:40.789913 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:45:41 crc kubenswrapper[4972]: I0228 11:45:41.977769 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerStarted","Data":"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b"} Feb 28 11:45:42 crc kubenswrapper[4972]: I0228 11:45:42.988562 4972 generic.go:334] "Generic (PLEG): container finished" podID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerID="30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b" exitCode=0 Feb 28 11:45:42 crc kubenswrapper[4972]: I0228 11:45:42.988614 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerDied","Data":"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b"} Feb 28 11:45:43 crc kubenswrapper[4972]: I0228 11:45:43.568311 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-mrgf8_74cd6520-9a2a-4e93-8536-2817d087ba65/cert-manager-controller/0.log" Feb 28 11:45:43 crc kubenswrapper[4972]: I0228 11:45:43.901967 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-76jg5_c0da0fed-5127-4b89-a79f-73e3d9c1ceee/cert-manager-webhook/0.log" Feb 28 11:45:43 crc kubenswrapper[4972]: I0228 11:45:43.940955 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-ptx2h_0dcc4f20-2ac3-4a75-8674-dfe9800d1844/cert-manager-cainjector/0.log" Feb 28 11:45:44 crc kubenswrapper[4972]: I0228 11:45:44.000617 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerStarted","Data":"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6"} Feb 28 11:45:44 crc kubenswrapper[4972]: I0228 11:45:44.026272 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bx5n6" podStartSLOduration=2.575166516 podStartE2EDuration="6.026255888s" podCreationTimestamp="2026-02-28 11:45:38 +0000 UTC" firstStartedPulling="2026-02-28 11:45:39.963110668 +0000 UTC m=+4196.875118396" lastFinishedPulling="2026-02-28 11:45:43.41420003 +0000 UTC m=+4200.326207768" observedRunningTime="2026-02-28 11:45:44.020295818 +0000 UTC m=+4200.932303576" watchObservedRunningTime="2026-02-28 11:45:44.026255888 +0000 UTC m=+4200.938263626" Feb 28 11:45:49 crc kubenswrapper[4972]: I0228 11:45:49.097545 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:49 crc kubenswrapper[4972]: I0228 11:45:49.098446 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:49 crc kubenswrapper[4972]: I0228 11:45:49.172021 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:50 crc kubenswrapper[4972]: I0228 11:45:50.139136 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:50 crc kubenswrapper[4972]: I0228 11:45:50.201906 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.092084 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bx5n6" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="registry-server" containerID="cri-o://34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6" gracePeriod=2 Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.585794 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.749274 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities\") pod \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.750039 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq2jn\" (UniqueName: \"kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn\") pod \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.750175 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content\") pod \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\" (UID: \"69ef38b6-d0d5-4712-a59c-6268b55a65f5\") " Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.750783 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities" (OuterVolumeSpecName: "utilities") pod "69ef38b6-d0d5-4712-a59c-6268b55a65f5" (UID: "69ef38b6-d0d5-4712-a59c-6268b55a65f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.758973 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn" (OuterVolumeSpecName: "kube-api-access-vq2jn") pod "69ef38b6-d0d5-4712-a59c-6268b55a65f5" (UID: "69ef38b6-d0d5-4712-a59c-6268b55a65f5"). InnerVolumeSpecName "kube-api-access-vq2jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.778232 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69ef38b6-d0d5-4712-a59c-6268b55a65f5" (UID: "69ef38b6-d0d5-4712-a59c-6268b55a65f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.853699 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.853814 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69ef38b6-d0d5-4712-a59c-6268b55a65f5-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:52 crc kubenswrapper[4972]: I0228 11:45:52.853867 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq2jn\" (UniqueName: \"kubernetes.io/projected/69ef38b6-d0d5-4712-a59c-6268b55a65f5-kube-api-access-vq2jn\") on node \"crc\" DevicePath \"\"" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.105100 4972 generic.go:334] "Generic (PLEG): container finished" podID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerID="34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6" exitCode=0 Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.105210 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx5n6" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.105230 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerDied","Data":"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6"} Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.105322 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx5n6" event={"ID":"69ef38b6-d0d5-4712-a59c-6268b55a65f5","Type":"ContainerDied","Data":"af424507ce75d4adbec1446182c3f99a59f6b51eb222dd333846951b6661ca56"} Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.105356 4972 scope.go:117] "RemoveContainer" containerID="34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.147141 4972 scope.go:117] "RemoveContainer" containerID="30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.159445 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.168605 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx5n6"] Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.170601 4972 scope.go:117] "RemoveContainer" containerID="5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.229066 4972 scope.go:117] "RemoveContainer" containerID="34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6" Feb 28 11:45:53 crc kubenswrapper[4972]: E0228 11:45:53.229671 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6\": container with ID starting with 34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6 not found: ID does not exist" containerID="34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.229714 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6"} err="failed to get container status \"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6\": rpc error: code = NotFound desc = could not find container \"34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6\": container with ID starting with 34ff12b0c188b4079dbfc8318f5ac41479a1cb6141c8c92280bd5e3b95b6c7f6 not found: ID does not exist" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.229743 4972 scope.go:117] "RemoveContainer" containerID="30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b" Feb 28 11:45:53 crc kubenswrapper[4972]: E0228 11:45:53.230264 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b\": container with ID starting with 30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b not found: ID does not exist" containerID="30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.230382 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b"} err="failed to get container status \"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b\": rpc error: code = NotFound desc = could not find container \"30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b\": container with ID starting with 30dc591860c35e55343640bb54099dd6088bce11b8a4fe6fc945d42cd956129b not found: ID does not exist" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.230504 4972 scope.go:117] "RemoveContainer" containerID="5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398" Feb 28 11:45:53 crc kubenswrapper[4972]: E0228 11:45:53.230914 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398\": container with ID starting with 5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398 not found: ID does not exist" containerID="5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.230942 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398"} err="failed to get container status \"5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398\": rpc error: code = NotFound desc = could not find container \"5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398\": container with ID starting with 5d7b65d842943d14f1a8e46e29d7b5744dccc64262957ac2806ef2d7f2c7f398 not found: ID does not exist" Feb 28 11:45:53 crc kubenswrapper[4972]: I0228 11:45:53.800536 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" path="/var/lib/kubelet/pods/69ef38b6-d0d5-4712-a59c-6268b55a65f5/volumes" Feb 28 11:45:55 crc kubenswrapper[4972]: I0228 11:45:55.789368 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:45:55 crc kubenswrapper[4972]: E0228 11:45:55.790027 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.167389 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537986-nssb7"] Feb 28 11:46:00 crc kubenswrapper[4972]: E0228 11:46:00.168632 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="registry-server" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.168652 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="registry-server" Feb 28 11:46:00 crc kubenswrapper[4972]: E0228 11:46:00.168713 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="extract-utilities" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.168726 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="extract-utilities" Feb 28 11:46:00 crc kubenswrapper[4972]: E0228 11:46:00.168742 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="extract-content" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.168752 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="extract-content" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.169038 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ef38b6-d0d5-4712-a59c-6268b55a65f5" containerName="registry-server" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.170071 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.172802 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.173070 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.174617 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.182395 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537986-nssb7"] Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.225832 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qwkk\" (UniqueName: \"kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk\") pod \"auto-csr-approver-29537986-nssb7\" (UID: \"dfe377c1-b766-4dc0-ab7a-6fa0c000d981\") " pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.328788 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qwkk\" (UniqueName: \"kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk\") pod \"auto-csr-approver-29537986-nssb7\" (UID: \"dfe377c1-b766-4dc0-ab7a-6fa0c000d981\") " pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.355994 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qwkk\" (UniqueName: \"kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk\") pod \"auto-csr-approver-29537986-nssb7\" (UID: \"dfe377c1-b766-4dc0-ab7a-6fa0c000d981\") " pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.530039 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:00 crc kubenswrapper[4972]: I0228 11:46:00.960289 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537986-nssb7"] Feb 28 11:46:01 crc kubenswrapper[4972]: I0228 11:46:01.228859 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537986-nssb7" event={"ID":"dfe377c1-b766-4dc0-ab7a-6fa0c000d981","Type":"ContainerStarted","Data":"ee86f18ad6a05d3d6bb474748a8484b313d0e4445c7052a017cd644a81a97a26"} Feb 28 11:46:01 crc kubenswrapper[4972]: I0228 11:46:01.585666 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-pdwzw_d5746037-ce0c-44c8-9874-23ac38ad6509/nmstate-console-plugin/0.log" Feb 28 11:46:01 crc kubenswrapper[4972]: I0228 11:46:01.799612 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-tt7f6_1013fa3e-26c1-4b74-a4f5-55c3b18fe34d/nmstate-handler/0.log" Feb 28 11:46:01 crc kubenswrapper[4972]: I0228 11:46:01.822120 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-sx977_a35c5570-d7ee-438e-8768-0daf418f5304/kube-rbac-proxy/0.log" Feb 28 11:46:01 crc kubenswrapper[4972]: I0228 11:46:01.938193 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-sx977_a35c5570-d7ee-438e-8768-0daf418f5304/nmstate-metrics/0.log" Feb 28 11:46:02 crc kubenswrapper[4972]: I0228 11:46:02.126395 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-srsp6_d8bdef5d-8a08-4e53-bb43-2f721ed3570e/nmstate-operator/0.log" Feb 28 11:46:02 crc kubenswrapper[4972]: I0228 11:46:02.167087 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-2z62s_b9a8aae0-4abb-4237-b2da-36542bac6c31/nmstate-webhook/0.log" Feb 28 11:46:02 crc kubenswrapper[4972]: I0228 11:46:02.238476 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537986-nssb7" event={"ID":"dfe377c1-b766-4dc0-ab7a-6fa0c000d981","Type":"ContainerStarted","Data":"49befc062660d9dd2995c4ee2f3fe57485d96db7627b49121861d1460c2dfad9"} Feb 28 11:46:02 crc kubenswrapper[4972]: I0228 11:46:02.253321 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29537986-nssb7" podStartSLOduration=1.348768722 podStartE2EDuration="2.25330535s" podCreationTimestamp="2026-02-28 11:46:00 +0000 UTC" firstStartedPulling="2026-02-28 11:46:00.962676896 +0000 UTC m=+4217.874684634" lastFinishedPulling="2026-02-28 11:46:01.867213524 +0000 UTC m=+4218.779221262" observedRunningTime="2026-02-28 11:46:02.25109355 +0000 UTC m=+4219.163101288" watchObservedRunningTime="2026-02-28 11:46:02.25330535 +0000 UTC m=+4219.165313088" Feb 28 11:46:03 crc kubenswrapper[4972]: I0228 11:46:03.250984 4972 generic.go:334] "Generic (PLEG): container finished" podID="dfe377c1-b766-4dc0-ab7a-6fa0c000d981" containerID="49befc062660d9dd2995c4ee2f3fe57485d96db7627b49121861d1460c2dfad9" exitCode=0 Feb 28 11:46:03 crc kubenswrapper[4972]: I0228 11:46:03.251049 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537986-nssb7" event={"ID":"dfe377c1-b766-4dc0-ab7a-6fa0c000d981","Type":"ContainerDied","Data":"49befc062660d9dd2995c4ee2f3fe57485d96db7627b49121861d1460c2dfad9"} Feb 28 11:46:04 crc kubenswrapper[4972]: I0228 11:46:04.703824 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:04 crc kubenswrapper[4972]: I0228 11:46:04.835551 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qwkk\" (UniqueName: \"kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk\") pod \"dfe377c1-b766-4dc0-ab7a-6fa0c000d981\" (UID: \"dfe377c1-b766-4dc0-ab7a-6fa0c000d981\") " Feb 28 11:46:04 crc kubenswrapper[4972]: I0228 11:46:04.846194 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk" (OuterVolumeSpecName: "kube-api-access-2qwkk") pod "dfe377c1-b766-4dc0-ab7a-6fa0c000d981" (UID: "dfe377c1-b766-4dc0-ab7a-6fa0c000d981"). InnerVolumeSpecName "kube-api-access-2qwkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:46:04 crc kubenswrapper[4972]: I0228 11:46:04.939552 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qwkk\" (UniqueName: \"kubernetes.io/projected/dfe377c1-b766-4dc0-ab7a-6fa0c000d981-kube-api-access-2qwkk\") on node \"crc\" DevicePath \"\"" Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.280328 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537986-nssb7" event={"ID":"dfe377c1-b766-4dc0-ab7a-6fa0c000d981","Type":"ContainerDied","Data":"ee86f18ad6a05d3d6bb474748a8484b313d0e4445c7052a017cd644a81a97a26"} Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.280389 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee86f18ad6a05d3d6bb474748a8484b313d0e4445c7052a017cd644a81a97a26" Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.280522 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537986-nssb7" Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.334518 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537980-qtngz"] Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.341932 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537980-qtngz"] Feb 28 11:46:05 crc kubenswrapper[4972]: I0228 11:46:05.805861 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0da5c8-255b-46b6-acf9-1cd155990486" path="/var/lib/kubelet/pods/2f0da5c8-255b-46b6-acf9-1cd155990486/volumes" Feb 28 11:46:06 crc kubenswrapper[4972]: I0228 11:46:06.788905 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:46:06 crc kubenswrapper[4972]: E0228 11:46:06.789354 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:46:16 crc kubenswrapper[4972]: I0228 11:46:16.471020 4972 scope.go:117] "RemoveContainer" containerID="5cba31be8163144c6fc906581293aa0166e437b5af21f5829f824c195ce499ba" Feb 28 11:46:18 crc kubenswrapper[4972]: I0228 11:46:18.789693 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:46:18 crc kubenswrapper[4972]: E0228 11:46:18.790300 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:46:30 crc kubenswrapper[4972]: I0228 11:46:30.790397 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:46:31 crc kubenswrapper[4972]: I0228 11:46:31.861113 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5"} Feb 28 11:46:37 crc kubenswrapper[4972]: I0228 11:46:37.626520 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-lxdv7_47d6aa51-c3aa-4f21-a02f-36663c003b7f/kube-rbac-proxy/0.log" Feb 28 11:46:37 crc kubenswrapper[4972]: I0228 11:46:37.749643 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-lxdv7_47d6aa51-c3aa-4f21-a02f-36663c003b7f/controller/0.log" Feb 28 11:46:37 crc kubenswrapper[4972]: I0228 11:46:37.865516 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.009847 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.057752 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.070578 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.098265 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.252503 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.259596 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.278356 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.334051 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.500591 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-reloader/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.531566 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/controller/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.560057 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-frr-files/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.566724 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/cp-metrics/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.732891 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/frr-metrics/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.793610 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/kube-rbac-proxy/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.815833 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/kube-rbac-proxy-frr/0.log" Feb 28 11:46:38 crc kubenswrapper[4972]: I0228 11:46:38.929841 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/reloader/0.log" Feb 28 11:46:39 crc kubenswrapper[4972]: I0228 11:46:39.752390 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-c9c7m_c4ddb606-f9be-489d-8c0b-5c89adbb265a/frr-k8s-webhook-server/0.log" Feb 28 11:46:39 crc kubenswrapper[4972]: I0228 11:46:39.763684 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57bd854678-2r552_7d4fb02a-0341-49cf-947f-86fb93ac710b/manager/0.log" Feb 28 11:46:40 crc kubenswrapper[4972]: I0228 11:46:40.022001 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5cdfc6748f-z4vkk_ac415cf4-e2bc-471a-91aa-2752b88b2636/webhook-server/0.log" Feb 28 11:46:40 crc kubenswrapper[4972]: I0228 11:46:40.262227 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n2xm6_b387ac1b-669a-43f3-b336-7df69f886a1a/kube-rbac-proxy/0.log" Feb 28 11:46:40 crc kubenswrapper[4972]: I0228 11:46:40.564207 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7n9h_0f19a4a9-5554-4e38-a29b-1d1ffde9fc87/frr/0.log" Feb 28 11:46:40 crc kubenswrapper[4972]: I0228 11:46:40.635965 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-n2xm6_b387ac1b-669a-43f3-b336-7df69f886a1a/speaker/0.log" Feb 28 11:46:56 crc kubenswrapper[4972]: I0228 11:46:56.829759 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:46:56 crc kubenswrapper[4972]: I0228 11:46:56.927329 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:46:56 crc kubenswrapper[4972]: I0228 11:46:56.974292 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.033328 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.162146 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/util/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.175471 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/pull/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.197223 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82blfnm_2eabfdfb-f2b5-458a-8f35-8c1aff548e91/extract/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.330671 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.487117 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.489587 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.511743 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.703072 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-utilities/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.715279 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/extract-content/0.log" Feb 28 11:46:57 crc kubenswrapper[4972]: I0228 11:46:57.878527 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.080506 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.120279 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2sr94_1bd5b0fb-5083-4eff-8225-02ba27a56628/registry-server/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.157957 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.197744 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.299945 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-utilities/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.319939 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/extract-content/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.548426 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.737603 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.753953 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.805927 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.903072 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jzl5b_c549bc93-54c0-4ead-95d0-3b94f86af486/registry-server/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.938519 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/util/0.log" Feb 28 11:46:58 crc kubenswrapper[4972]: I0228 11:46:58.974026 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/pull/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.054000 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4cnrgx_38feb8ea-eefb-4037-ba7a-b974e35ea6c3/extract/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.126254 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-m88zk_72e5fba8-6a6e-4310-bcb6-ad064ec74bb9/marketplace-operator/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.223415 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.398324 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.417441 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.433084 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.608868 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-utilities/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.650679 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/extract-content/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.730551 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ggdk_0c340c42-4bb5-4395-a1cf-2cbdcd29f894/registry-server/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.837996 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.965203 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:46:59 crc kubenswrapper[4972]: I0228 11:46:59.981637 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:47:00 crc kubenswrapper[4972]: I0228 11:47:00.022769 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:47:00 crc kubenswrapper[4972]: I0228 11:47:00.706112 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-utilities/0.log" Feb 28 11:47:00 crc kubenswrapper[4972]: I0228 11:47:00.757496 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/extract-content/0.log" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.173434 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-825kc_0d7d47c5-06cc-48b9-b7f4-836f7bffc194/registry-server/0.log" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.734761 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:01 crc kubenswrapper[4972]: E0228 11:47:01.735269 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe377c1-b766-4dc0-ab7a-6fa0c000d981" containerName="oc" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.735294 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe377c1-b766-4dc0-ab7a-6fa0c000d981" containerName="oc" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.735543 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe377c1-b766-4dc0-ab7a-6fa0c000d981" containerName="oc" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.737208 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.756916 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.820842 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgmcm\" (UniqueName: \"kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.820958 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.821007 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.922740 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgmcm\" (UniqueName: \"kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.922839 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.922884 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.923373 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:01 crc kubenswrapper[4972]: I0228 11:47:01.923530 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:02 crc kubenswrapper[4972]: I0228 11:47:02.515388 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgmcm\" (UniqueName: \"kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm\") pod \"redhat-operators-7whgf\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:02 crc kubenswrapper[4972]: I0228 11:47:02.699249 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:03 crc kubenswrapper[4972]: I0228 11:47:03.219598 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:04 crc kubenswrapper[4972]: I0228 11:47:04.208749 4972 generic.go:334] "Generic (PLEG): container finished" podID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerID="2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773" exitCode=0 Feb 28 11:47:04 crc kubenswrapper[4972]: I0228 11:47:04.208827 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerDied","Data":"2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773"} Feb 28 11:47:04 crc kubenswrapper[4972]: I0228 11:47:04.209348 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerStarted","Data":"18fb4d8fe46bfa7ce27e145ebe4d564edf4fc502bcd9b3726f94ff7e4c36e35b"} Feb 28 11:47:05 crc kubenswrapper[4972]: I0228 11:47:05.226865 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerStarted","Data":"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5"} Feb 28 11:47:09 crc kubenswrapper[4972]: I0228 11:47:09.275423 4972 generic.go:334] "Generic (PLEG): container finished" podID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerID="1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5" exitCode=0 Feb 28 11:47:09 crc kubenswrapper[4972]: I0228 11:47:09.275509 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerDied","Data":"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5"} Feb 28 11:47:10 crc kubenswrapper[4972]: I0228 11:47:10.287988 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerStarted","Data":"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377"} Feb 28 11:47:10 crc kubenswrapper[4972]: I0228 11:47:10.305784 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7whgf" podStartSLOduration=3.821225579 podStartE2EDuration="9.305766305s" podCreationTimestamp="2026-02-28 11:47:01 +0000 UTC" firstStartedPulling="2026-02-28 11:47:04.211531744 +0000 UTC m=+4281.123539522" lastFinishedPulling="2026-02-28 11:47:09.69607247 +0000 UTC m=+4286.608080248" observedRunningTime="2026-02-28 11:47:10.30409281 +0000 UTC m=+4287.216100548" watchObservedRunningTime="2026-02-28 11:47:10.305766305 +0000 UTC m=+4287.217774043" Feb 28 11:47:12 crc kubenswrapper[4972]: I0228 11:47:12.700228 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:12 crc kubenswrapper[4972]: I0228 11:47:12.700649 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:13 crc kubenswrapper[4972]: I0228 11:47:13.780170 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7whgf" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" probeResult="failure" output=< Feb 28 11:47:13 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:47:13 crc kubenswrapper[4972]: > Feb 28 11:47:23 crc kubenswrapper[4972]: I0228 11:47:23.752231 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7whgf" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" probeResult="failure" output=< Feb 28 11:47:23 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:47:23 crc kubenswrapper[4972]: > Feb 28 11:47:33 crc kubenswrapper[4972]: I0228 11:47:33.765703 4972 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7whgf" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" probeResult="failure" output=< Feb 28 11:47:33 crc kubenswrapper[4972]: timeout: failed to connect service ":50051" within 1s Feb 28 11:47:33 crc kubenswrapper[4972]: > Feb 28 11:47:36 crc kubenswrapper[4972]: E0228 11:47:36.881350 4972 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:55230->38.102.83.12:44997: write tcp 38.102.83.12:55230->38.102.83.12:44997: write: broken pipe Feb 28 11:47:42 crc kubenswrapper[4972]: I0228 11:47:42.773650 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:42 crc kubenswrapper[4972]: I0228 11:47:42.859511 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:43 crc kubenswrapper[4972]: I0228 11:47:43.054949 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:44 crc kubenswrapper[4972]: I0228 11:47:44.615498 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7whgf" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" containerID="cri-o://e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377" gracePeriod=2 Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.111617 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.176714 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgmcm\" (UniqueName: \"kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm\") pod \"89a9456c-c7da-44ff-a561-8c81cfba8842\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.176808 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content\") pod \"89a9456c-c7da-44ff-a561-8c81cfba8842\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.176890 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities\") pod \"89a9456c-c7da-44ff-a561-8c81cfba8842\" (UID: \"89a9456c-c7da-44ff-a561-8c81cfba8842\") " Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.178820 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities" (OuterVolumeSpecName: "utilities") pod "89a9456c-c7da-44ff-a561-8c81cfba8842" (UID: "89a9456c-c7da-44ff-a561-8c81cfba8842"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.192634 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm" (OuterVolumeSpecName: "kube-api-access-lgmcm") pod "89a9456c-c7da-44ff-a561-8c81cfba8842" (UID: "89a9456c-c7da-44ff-a561-8c81cfba8842"). InnerVolumeSpecName "kube-api-access-lgmcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.279756 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgmcm\" (UniqueName: \"kubernetes.io/projected/89a9456c-c7da-44ff-a561-8c81cfba8842-kube-api-access-lgmcm\") on node \"crc\" DevicePath \"\"" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.280035 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.307570 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89a9456c-c7da-44ff-a561-8c81cfba8842" (UID: "89a9456c-c7da-44ff-a561-8c81cfba8842"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.381420 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a9456c-c7da-44ff-a561-8c81cfba8842-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.628523 4972 generic.go:334] "Generic (PLEG): container finished" podID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerID="e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377" exitCode=0 Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.628590 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerDied","Data":"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377"} Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.628637 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7whgf" event={"ID":"89a9456c-c7da-44ff-a561-8c81cfba8842","Type":"ContainerDied","Data":"18fb4d8fe46bfa7ce27e145ebe4d564edf4fc502bcd9b3726f94ff7e4c36e35b"} Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.628666 4972 scope.go:117] "RemoveContainer" containerID="e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.628856 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7whgf" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.676926 4972 scope.go:117] "RemoveContainer" containerID="1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.724264 4972 scope.go:117] "RemoveContainer" containerID="2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773" Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.729817 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.756553 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7whgf"] Feb 28 11:47:45 crc kubenswrapper[4972]: I0228 11:47:45.803852 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" path="/var/lib/kubelet/pods/89a9456c-c7da-44ff-a561-8c81cfba8842/volumes" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.494102 4972 scope.go:117] "RemoveContainer" containerID="e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377" Feb 28 11:47:46 crc kubenswrapper[4972]: E0228 11:47:46.495054 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377\": container with ID starting with e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377 not found: ID does not exist" containerID="e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.495110 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377"} err="failed to get container status \"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377\": rpc error: code = NotFound desc = could not find container \"e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377\": container with ID starting with e450cb5e14c9a09790fb58d28ce2d7c4997f3e0c46fdbf4f12e8c3c3d928e377 not found: ID does not exist" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.495144 4972 scope.go:117] "RemoveContainer" containerID="1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5" Feb 28 11:47:46 crc kubenswrapper[4972]: E0228 11:47:46.495650 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5\": container with ID starting with 1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5 not found: ID does not exist" containerID="1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.495798 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5"} err="failed to get container status \"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5\": rpc error: code = NotFound desc = could not find container \"1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5\": container with ID starting with 1287c494fa423cf1946358f60ffdda82c4e221c8b47a453c2b9acf9ba2c6f3d5 not found: ID does not exist" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.495832 4972 scope.go:117] "RemoveContainer" containerID="2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773" Feb 28 11:47:46 crc kubenswrapper[4972]: E0228 11:47:46.496204 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773\": container with ID starting with 2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773 not found: ID does not exist" containerID="2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773" Feb 28 11:47:46 crc kubenswrapper[4972]: I0228 11:47:46.496247 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773"} err="failed to get container status \"2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773\": rpc error: code = NotFound desc = could not find container \"2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773\": container with ID starting with 2c18ab86304db52dc8aae9c334cb9bbb44652971fb9cacdbfd8810418c9d9773 not found: ID does not exist" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.141959 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537988-hd8th"] Feb 28 11:48:00 crc kubenswrapper[4972]: E0228 11:48:00.143095 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="extract-content" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.143111 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="extract-content" Feb 28 11:48:00 crc kubenswrapper[4972]: E0228 11:48:00.143137 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.143145 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" Feb 28 11:48:00 crc kubenswrapper[4972]: E0228 11:48:00.143166 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="extract-utilities" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.143174 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="extract-utilities" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.143537 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a9456c-c7da-44ff-a561-8c81cfba8842" containerName="registry-server" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.144348 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.146542 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.146985 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.147727 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.150662 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537988-hd8th"] Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.214143 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhnc8\" (UniqueName: \"kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8\") pod \"auto-csr-approver-29537988-hd8th\" (UID: \"4b420004-06f7-42ff-99f1-de2406811963\") " pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.316170 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhnc8\" (UniqueName: \"kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8\") pod \"auto-csr-approver-29537988-hd8th\" (UID: \"4b420004-06f7-42ff-99f1-de2406811963\") " pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.348439 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhnc8\" (UniqueName: \"kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8\") pod \"auto-csr-approver-29537988-hd8th\" (UID: \"4b420004-06f7-42ff-99f1-de2406811963\") " pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:00 crc kubenswrapper[4972]: I0228 11:48:00.468904 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:01 crc kubenswrapper[4972]: I0228 11:48:01.535834 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537988-hd8th"] Feb 28 11:48:01 crc kubenswrapper[4972]: I0228 11:48:01.830691 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537988-hd8th" event={"ID":"4b420004-06f7-42ff-99f1-de2406811963","Type":"ContainerStarted","Data":"bdc0f3e70cf467cd7ecacd0546374cd3e6122ad8606916ccd80d710748dd69f5"} Feb 28 11:48:03 crc kubenswrapper[4972]: I0228 11:48:03.855583 4972 generic.go:334] "Generic (PLEG): container finished" podID="4b420004-06f7-42ff-99f1-de2406811963" containerID="79b83a66d79cb322016cf32e7b3b9314bb74b28b610d5a7edceae1c41fba9810" exitCode=0 Feb 28 11:48:03 crc kubenswrapper[4972]: I0228 11:48:03.855667 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537988-hd8th" event={"ID":"4b420004-06f7-42ff-99f1-de2406811963","Type":"ContainerDied","Data":"79b83a66d79cb322016cf32e7b3b9314bb74b28b610d5a7edceae1c41fba9810"} Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.331569 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.423253 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhnc8\" (UniqueName: \"kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8\") pod \"4b420004-06f7-42ff-99f1-de2406811963\" (UID: \"4b420004-06f7-42ff-99f1-de2406811963\") " Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.431794 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8" (OuterVolumeSpecName: "kube-api-access-xhnc8") pod "4b420004-06f7-42ff-99f1-de2406811963" (UID: "4b420004-06f7-42ff-99f1-de2406811963"). InnerVolumeSpecName "kube-api-access-xhnc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.526444 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhnc8\" (UniqueName: \"kubernetes.io/projected/4b420004-06f7-42ff-99f1-de2406811963-kube-api-access-xhnc8\") on node \"crc\" DevicePath \"\"" Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.920433 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537988-hd8th" event={"ID":"4b420004-06f7-42ff-99f1-de2406811963","Type":"ContainerDied","Data":"bdc0f3e70cf467cd7ecacd0546374cd3e6122ad8606916ccd80d710748dd69f5"} Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.920509 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537988-hd8th" Feb 28 11:48:05 crc kubenswrapper[4972]: I0228 11:48:05.920528 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdc0f3e70cf467cd7ecacd0546374cd3e6122ad8606916ccd80d710748dd69f5" Feb 28 11:48:06 crc kubenswrapper[4972]: I0228 11:48:06.423816 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537982-kgn58"] Feb 28 11:48:06 crc kubenswrapper[4972]: I0228 11:48:06.432140 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537982-kgn58"] Feb 28 11:48:07 crc kubenswrapper[4972]: I0228 11:48:07.803133 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b7ff79-c7ca-48bf-a36a-210158b98164" path="/var/lib/kubelet/pods/65b7ff79-c7ca-48bf-a36a-210158b98164/volumes" Feb 28 11:48:16 crc kubenswrapper[4972]: I0228 11:48:16.646220 4972 scope.go:117] "RemoveContainer" containerID="0ca00e3e9aa68e5a7d16b43c9f6d72fc3c0d8b50975ad7e00c70aaf4c6e73707" Feb 28 11:48:51 crc kubenswrapper[4972]: I0228 11:48:51.441152 4972 generic.go:334] "Generic (PLEG): container finished" podID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerID="1987bcb7ecbe11c4d7e7920f5a3587354b67f7adcef516f405313a7edc82b515" exitCode=0 Feb 28 11:48:51 crc kubenswrapper[4972]: I0228 11:48:51.441283 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v89xz/must-gather-zb8lh" event={"ID":"38760a3b-d6fe-466f-b10a-c2d3336082ae","Type":"ContainerDied","Data":"1987bcb7ecbe11c4d7e7920f5a3587354b67f7adcef516f405313a7edc82b515"} Feb 28 11:48:51 crc kubenswrapper[4972]: I0228 11:48:51.443090 4972 scope.go:117] "RemoveContainer" containerID="1987bcb7ecbe11c4d7e7920f5a3587354b67f7adcef516f405313a7edc82b515" Feb 28 11:48:52 crc kubenswrapper[4972]: I0228 11:48:52.531253 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v89xz_must-gather-zb8lh_38760a3b-d6fe-466f-b10a-c2d3336082ae/gather/0.log" Feb 28 11:48:58 crc kubenswrapper[4972]: I0228 11:48:58.890919 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:48:58 crc kubenswrapper[4972]: I0228 11:48:58.891651 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.346613 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v89xz/must-gather-zb8lh"] Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.347993 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-v89xz/must-gather-zb8lh" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="copy" containerID="cri-o://9cc4879980ca1f8e9275270420781ebf6fade9b4c62c380ce26c1a2a2d5aabbd" gracePeriod=2 Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.356182 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v89xz/must-gather-zb8lh"] Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.572055 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v89xz_must-gather-zb8lh_38760a3b-d6fe-466f-b10a-c2d3336082ae/copy/0.log" Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.572440 4972 generic.go:334] "Generic (PLEG): container finished" podID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerID="9cc4879980ca1f8e9275270420781ebf6fade9b4c62c380ce26c1a2a2d5aabbd" exitCode=143 Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.767207 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v89xz_must-gather-zb8lh_38760a3b-d6fe-466f-b10a-c2d3336082ae/copy/0.log" Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.767997 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.880768 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d8th\" (UniqueName: \"kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th\") pod \"38760a3b-d6fe-466f-b10a-c2d3336082ae\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.880807 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output\") pod \"38760a3b-d6fe-466f-b10a-c2d3336082ae\" (UID: \"38760a3b-d6fe-466f-b10a-c2d3336082ae\") " Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.890044 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th" (OuterVolumeSpecName: "kube-api-access-7d8th") pod "38760a3b-d6fe-466f-b10a-c2d3336082ae" (UID: "38760a3b-d6fe-466f-b10a-c2d3336082ae"). InnerVolumeSpecName "kube-api-access-7d8th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:49:02 crc kubenswrapper[4972]: I0228 11:49:02.983825 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d8th\" (UniqueName: \"kubernetes.io/projected/38760a3b-d6fe-466f-b10a-c2d3336082ae-kube-api-access-7d8th\") on node \"crc\" DevicePath \"\"" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.022990 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "38760a3b-d6fe-466f-b10a-c2d3336082ae" (UID: "38760a3b-d6fe-466f-b10a-c2d3336082ae"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.086184 4972 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38760a3b-d6fe-466f-b10a-c2d3336082ae-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.582526 4972 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v89xz_must-gather-zb8lh_38760a3b-d6fe-466f-b10a-c2d3336082ae/copy/0.log" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.583916 4972 scope.go:117] "RemoveContainer" containerID="9cc4879980ca1f8e9275270420781ebf6fade9b4c62c380ce26c1a2a2d5aabbd" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.583984 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v89xz/must-gather-zb8lh" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.626258 4972 scope.go:117] "RemoveContainer" containerID="1987bcb7ecbe11c4d7e7920f5a3587354b67f7adcef516f405313a7edc82b515" Feb 28 11:49:03 crc kubenswrapper[4972]: I0228 11:49:03.802597 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" path="/var/lib/kubelet/pods/38760a3b-d6fe-466f-b10a-c2d3336082ae/volumes" Feb 28 11:49:16 crc kubenswrapper[4972]: I0228 11:49:16.826421 4972 scope.go:117] "RemoveContainer" containerID="33dfba4f2daffd9b993abb3bb646656574b45aded0d0dd56b678c748c7ac833f" Feb 28 11:49:28 crc kubenswrapper[4972]: I0228 11:49:28.890621 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:49:28 crc kubenswrapper[4972]: I0228 11:49:28.891306 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:49:58 crc kubenswrapper[4972]: I0228 11:49:58.891383 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:49:58 crc kubenswrapper[4972]: I0228 11:49:58.892166 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:49:58 crc kubenswrapper[4972]: I0228 11:49:58.892236 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:49:58 crc kubenswrapper[4972]: I0228 11:49:58.893382 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:49:58 crc kubenswrapper[4972]: I0228 11:49:58.893534 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5" gracePeriod=600 Feb 28 11:49:59 crc kubenswrapper[4972]: I0228 11:49:59.332299 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5" exitCode=0 Feb 28 11:49:59 crc kubenswrapper[4972]: I0228 11:49:59.332400 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5"} Feb 28 11:49:59 crc kubenswrapper[4972]: I0228 11:49:59.332694 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerStarted","Data":"3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893"} Feb 28 11:49:59 crc kubenswrapper[4972]: I0228 11:49:59.332721 4972 scope.go:117] "RemoveContainer" containerID="5bdab33504a2103ee2d8b0c055cdbed829dc738b1a182a558c9213f9e6f5d4b3" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.157110 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537990-6njpt"] Feb 28 11:50:00 crc kubenswrapper[4972]: E0228 11:50:00.158401 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="copy" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.158423 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="copy" Feb 28 11:50:00 crc kubenswrapper[4972]: E0228 11:50:00.158447 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="gather" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.158640 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="gather" Feb 28 11:50:00 crc kubenswrapper[4972]: E0228 11:50:00.158685 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b420004-06f7-42ff-99f1-de2406811963" containerName="oc" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.158699 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b420004-06f7-42ff-99f1-de2406811963" containerName="oc" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.159028 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="gather" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.159055 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="38760a3b-d6fe-466f-b10a-c2d3336082ae" containerName="copy" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.159088 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b420004-06f7-42ff-99f1-de2406811963" containerName="oc" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.160168 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.164678 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.167189 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.176565 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.183001 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537990-6njpt"] Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.282256 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghldt\" (UniqueName: \"kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt\") pod \"auto-csr-approver-29537990-6njpt\" (UID: \"e4fbe8d1-5545-4f09-9617-cfccd430478e\") " pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.384445 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghldt\" (UniqueName: \"kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt\") pod \"auto-csr-approver-29537990-6njpt\" (UID: \"e4fbe8d1-5545-4f09-9617-cfccd430478e\") " pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.411512 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghldt\" (UniqueName: \"kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt\") pod \"auto-csr-approver-29537990-6njpt\" (UID: \"e4fbe8d1-5545-4f09-9617-cfccd430478e\") " pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.491234 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:00 crc kubenswrapper[4972]: I0228 11:50:00.993378 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537990-6njpt"] Feb 28 11:50:00 crc kubenswrapper[4972]: W0228 11:50:00.999706 4972 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4fbe8d1_5545_4f09_9617_cfccd430478e.slice/crio-3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166 WatchSource:0}: Error finding container 3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166: Status 404 returned error can't find the container with id 3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166 Feb 28 11:50:01 crc kubenswrapper[4972]: I0228 11:50:01.003150 4972 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 28 11:50:01 crc kubenswrapper[4972]: I0228 11:50:01.360242 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537990-6njpt" event={"ID":"e4fbe8d1-5545-4f09-9617-cfccd430478e","Type":"ContainerStarted","Data":"3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166"} Feb 28 11:50:03 crc kubenswrapper[4972]: I0228 11:50:03.391904 4972 generic.go:334] "Generic (PLEG): container finished" podID="e4fbe8d1-5545-4f09-9617-cfccd430478e" containerID="5fd003d5c154a884774dfd5d20e4727183e165f4e321adb214f1f8a8e2a1b8d9" exitCode=0 Feb 28 11:50:03 crc kubenswrapper[4972]: I0228 11:50:03.392036 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537990-6njpt" event={"ID":"e4fbe8d1-5545-4f09-9617-cfccd430478e","Type":"ContainerDied","Data":"5fd003d5c154a884774dfd5d20e4727183e165f4e321adb214f1f8a8e2a1b8d9"} Feb 28 11:50:04 crc kubenswrapper[4972]: I0228 11:50:04.837157 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.003987 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghldt\" (UniqueName: \"kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt\") pod \"e4fbe8d1-5545-4f09-9617-cfccd430478e\" (UID: \"e4fbe8d1-5545-4f09-9617-cfccd430478e\") " Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.012062 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt" (OuterVolumeSpecName: "kube-api-access-ghldt") pod "e4fbe8d1-5545-4f09-9617-cfccd430478e" (UID: "e4fbe8d1-5545-4f09-9617-cfccd430478e"). InnerVolumeSpecName "kube-api-access-ghldt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.107437 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghldt\" (UniqueName: \"kubernetes.io/projected/e4fbe8d1-5545-4f09-9617-cfccd430478e-kube-api-access-ghldt\") on node \"crc\" DevicePath \"\"" Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.419603 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537990-6njpt" event={"ID":"e4fbe8d1-5545-4f09-9617-cfccd430478e","Type":"ContainerDied","Data":"3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166"} Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.419643 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3108c43091e609bce98d8c26ccee56a74d9f513a953652af2f89837e644d5166" Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.419740 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537990-6njpt" Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.930547 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537984-gpnsb"] Feb 28 11:50:05 crc kubenswrapper[4972]: I0228 11:50:05.941038 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537984-gpnsb"] Feb 28 11:50:07 crc kubenswrapper[4972]: I0228 11:50:07.818653 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef06ae10-f323-4fb0-b44c-0a8f549e4915" path="/var/lib/kubelet/pods/ef06ae10-f323-4fb0-b44c-0a8f549e4915/volumes" Feb 28 11:50:16 crc kubenswrapper[4972]: I0228 11:50:16.879910 4972 scope.go:117] "RemoveContainer" containerID="d2e67657f393e73b1281dcacd11e5a9f4e97c47e6bf3bc6b1ec3bcb6c0b39a4f" Feb 28 11:50:27 crc kubenswrapper[4972]: I0228 11:50:27.715319 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="44e7c813-beaa-4015-a581-afcaea85077c" containerName="galera" probeResult="failure" output="command timed out" Feb 28 11:50:27 crc kubenswrapper[4972]: I0228 11:50:27.716705 4972 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="44e7c813-beaa-4015-a581-afcaea85077c" containerName="galera" probeResult="failure" output="command timed out" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.127074 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:42 crc kubenswrapper[4972]: E0228 11:51:42.128078 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4fbe8d1-5545-4f09-9617-cfccd430478e" containerName="oc" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.128094 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4fbe8d1-5545-4f09-9617-cfccd430478e" containerName="oc" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.128332 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4fbe8d1-5545-4f09-9617-cfccd430478e" containerName="oc" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.130413 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.140534 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.219601 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.220290 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.220412 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db2th\" (UniqueName: \"kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.321627 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.321984 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db2th\" (UniqueName: \"kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.322352 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.322834 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.323274 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.347494 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db2th\" (UniqueName: \"kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th\") pod \"community-operators-5mvnp\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:42 crc kubenswrapper[4972]: I0228 11:51:42.465298 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:43 crc kubenswrapper[4972]: I0228 11:51:43.077544 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:43 crc kubenswrapper[4972]: I0228 11:51:43.575240 4972 generic.go:334] "Generic (PLEG): container finished" podID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerID="632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9" exitCode=0 Feb 28 11:51:43 crc kubenswrapper[4972]: I0228 11:51:43.575310 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerDied","Data":"632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9"} Feb 28 11:51:43 crc kubenswrapper[4972]: I0228 11:51:43.575775 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerStarted","Data":"4e981d867555126ea00c9a71e006c8674fef12b96400e854fe3e17a00732a335"} Feb 28 11:51:44 crc kubenswrapper[4972]: I0228 11:51:44.588836 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerStarted","Data":"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b"} Feb 28 11:51:46 crc kubenswrapper[4972]: I0228 11:51:46.613105 4972 generic.go:334] "Generic (PLEG): container finished" podID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerID="6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b" exitCode=0 Feb 28 11:51:46 crc kubenswrapper[4972]: I0228 11:51:46.613203 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerDied","Data":"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b"} Feb 28 11:51:47 crc kubenswrapper[4972]: I0228 11:51:47.626112 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerStarted","Data":"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583"} Feb 28 11:51:47 crc kubenswrapper[4972]: I0228 11:51:47.653349 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5mvnp" podStartSLOduration=2.2142507240000002 podStartE2EDuration="5.653324475s" podCreationTimestamp="2026-02-28 11:51:42 +0000 UTC" firstStartedPulling="2026-02-28 11:51:43.577872583 +0000 UTC m=+4560.489880321" lastFinishedPulling="2026-02-28 11:51:47.016946324 +0000 UTC m=+4563.928954072" observedRunningTime="2026-02-28 11:51:47.644600071 +0000 UTC m=+4564.556607809" watchObservedRunningTime="2026-02-28 11:51:47.653324475 +0000 UTC m=+4564.565332253" Feb 28 11:51:52 crc kubenswrapper[4972]: I0228 11:51:52.465917 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:52 crc kubenswrapper[4972]: I0228 11:51:52.466856 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:52 crc kubenswrapper[4972]: I0228 11:51:52.540124 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:52 crc kubenswrapper[4972]: I0228 11:51:52.759145 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:52 crc kubenswrapper[4972]: I0228 11:51:52.837329 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:54 crc kubenswrapper[4972]: I0228 11:51:54.705150 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5mvnp" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="registry-server" containerID="cri-o://172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583" gracePeriod=2 Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.284287 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.400023 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content\") pod \"e074b90a-a99d-4903-abc3-fef3be3fe87f\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.400082 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities\") pod \"e074b90a-a99d-4903-abc3-fef3be3fe87f\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.400117 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db2th\" (UniqueName: \"kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th\") pod \"e074b90a-a99d-4903-abc3-fef3be3fe87f\" (UID: \"e074b90a-a99d-4903-abc3-fef3be3fe87f\") " Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.402809 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities" (OuterVolumeSpecName: "utilities") pod "e074b90a-a99d-4903-abc3-fef3be3fe87f" (UID: "e074b90a-a99d-4903-abc3-fef3be3fe87f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.406882 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th" (OuterVolumeSpecName: "kube-api-access-db2th") pod "e074b90a-a99d-4903-abc3-fef3be3fe87f" (UID: "e074b90a-a99d-4903-abc3-fef3be3fe87f"). InnerVolumeSpecName "kube-api-access-db2th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.468611 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e074b90a-a99d-4903-abc3-fef3be3fe87f" (UID: "e074b90a-a99d-4903-abc3-fef3be3fe87f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.502381 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.502413 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e074b90a-a99d-4903-abc3-fef3be3fe87f-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.502424 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db2th\" (UniqueName: \"kubernetes.io/projected/e074b90a-a99d-4903-abc3-fef3be3fe87f-kube-api-access-db2th\") on node \"crc\" DevicePath \"\"" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.715911 4972 generic.go:334] "Generic (PLEG): container finished" podID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerID="172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583" exitCode=0 Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.715987 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerDied","Data":"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583"} Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.716101 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mvnp" event={"ID":"e074b90a-a99d-4903-abc3-fef3be3fe87f","Type":"ContainerDied","Data":"4e981d867555126ea00c9a71e006c8674fef12b96400e854fe3e17a00732a335"} Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.716009 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mvnp" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.716134 4972 scope.go:117] "RemoveContainer" containerID="172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.755542 4972 scope.go:117] "RemoveContainer" containerID="6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.775649 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.776815 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5mvnp"] Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.782344 4972 scope.go:117] "RemoveContainer" containerID="632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.815431 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" path="/var/lib/kubelet/pods/e074b90a-a99d-4903-abc3-fef3be3fe87f/volumes" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853002 4972 scope.go:117] "RemoveContainer" containerID="172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583" Feb 28 11:51:55 crc kubenswrapper[4972]: E0228 11:51:55.853442 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583\": container with ID starting with 172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583 not found: ID does not exist" containerID="172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853484 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583"} err="failed to get container status \"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583\": rpc error: code = NotFound desc = could not find container \"172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583\": container with ID starting with 172ba95ff8e13998ad2f859535767cbb4628db3b5546f51818aef18da8541583 not found: ID does not exist" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853502 4972 scope.go:117] "RemoveContainer" containerID="6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b" Feb 28 11:51:55 crc kubenswrapper[4972]: E0228 11:51:55.853790 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b\": container with ID starting with 6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b not found: ID does not exist" containerID="6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853806 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b"} err="failed to get container status \"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b\": rpc error: code = NotFound desc = could not find container \"6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b\": container with ID starting with 6ba976021d2b0e15dece41443f4a359ca494e44ad06d5b0bf764605f5011009b not found: ID does not exist" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853819 4972 scope.go:117] "RemoveContainer" containerID="632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9" Feb 28 11:51:55 crc kubenswrapper[4972]: E0228 11:51:55.853974 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9\": container with ID starting with 632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9 not found: ID does not exist" containerID="632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9" Feb 28 11:51:55 crc kubenswrapper[4972]: I0228 11:51:55.853988 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9"} err="failed to get container status \"632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9\": rpc error: code = NotFound desc = could not find container \"632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9\": container with ID starting with 632e35fb877180c9984edd4e1d660e653fa39e6795f83fb2982e3fab4d014bf9 not found: ID does not exist" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.168309 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29537992-bkqjv"] Feb 28 11:52:00 crc kubenswrapper[4972]: E0228 11:52:00.169916 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="registry-server" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.169952 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="registry-server" Feb 28 11:52:00 crc kubenswrapper[4972]: E0228 11:52:00.169974 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="extract-content" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.169991 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="extract-content" Feb 28 11:52:00 crc kubenswrapper[4972]: E0228 11:52:00.170069 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="extract-utilities" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.170088 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="extract-utilities" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.170500 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="e074b90a-a99d-4903-abc3-fef3be3fe87f" containerName="registry-server" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.171572 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.174570 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.174841 4972 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.174857 4972 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-brkss" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.242923 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537992-bkqjv"] Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.340689 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpqgg\" (UniqueName: \"kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg\") pod \"auto-csr-approver-29537992-bkqjv\" (UID: \"ecfa7738-9ed8-475f-863e-ea582e7c1b8b\") " pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.442575 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpqgg\" (UniqueName: \"kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg\") pod \"auto-csr-approver-29537992-bkqjv\" (UID: \"ecfa7738-9ed8-475f-863e-ea582e7c1b8b\") " pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.485274 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpqgg\" (UniqueName: \"kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg\") pod \"auto-csr-approver-29537992-bkqjv\" (UID: \"ecfa7738-9ed8-475f-863e-ea582e7c1b8b\") " pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:00 crc kubenswrapper[4972]: I0228 11:52:00.567433 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:01 crc kubenswrapper[4972]: I0228 11:52:01.013114 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29537992-bkqjv"] Feb 28 11:52:01 crc kubenswrapper[4972]: I0228 11:52:01.810194 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" event={"ID":"ecfa7738-9ed8-475f-863e-ea582e7c1b8b","Type":"ContainerStarted","Data":"079772bd483dbfb3cc147485ee06973316451edc78aa41dcde0dea82fca22c18"} Feb 28 11:52:02 crc kubenswrapper[4972]: I0228 11:52:02.822268 4972 generic.go:334] "Generic (PLEG): container finished" podID="ecfa7738-9ed8-475f-863e-ea582e7c1b8b" containerID="f336ddf0e94eaf7ac14da6ee6c6e95c513d80f1f79bc60db3fbcbbc45c6ab763" exitCode=0 Feb 28 11:52:02 crc kubenswrapper[4972]: I0228 11:52:02.822434 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" event={"ID":"ecfa7738-9ed8-475f-863e-ea582e7c1b8b","Type":"ContainerDied","Data":"f336ddf0e94eaf7ac14da6ee6c6e95c513d80f1f79bc60db3fbcbbc45c6ab763"} Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.234573 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.429956 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpqgg\" (UniqueName: \"kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg\") pod \"ecfa7738-9ed8-475f-863e-ea582e7c1b8b\" (UID: \"ecfa7738-9ed8-475f-863e-ea582e7c1b8b\") " Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.436752 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg" (OuterVolumeSpecName: "kube-api-access-cpqgg") pod "ecfa7738-9ed8-475f-863e-ea582e7c1b8b" (UID: "ecfa7738-9ed8-475f-863e-ea582e7c1b8b"). InnerVolumeSpecName "kube-api-access-cpqgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.533864 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpqgg\" (UniqueName: \"kubernetes.io/projected/ecfa7738-9ed8-475f-863e-ea582e7c1b8b-kube-api-access-cpqgg\") on node \"crc\" DevicePath \"\"" Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.845846 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" event={"ID":"ecfa7738-9ed8-475f-863e-ea582e7c1b8b","Type":"ContainerDied","Data":"079772bd483dbfb3cc147485ee06973316451edc78aa41dcde0dea82fca22c18"} Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.845904 4972 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079772bd483dbfb3cc147485ee06973316451edc78aa41dcde0dea82fca22c18" Feb 28 11:52:04 crc kubenswrapper[4972]: I0228 11:52:04.845954 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29537992-bkqjv" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.346536 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29537986-nssb7"] Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.359747 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29537986-nssb7"] Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.808364 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe377c1-b766-4dc0-ab7a-6fa0c000d981" path="/var/lib/kubelet/pods/dfe377c1-b766-4dc0-ab7a-6fa0c000d981/volumes" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.875506 4972 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:05 crc kubenswrapper[4972]: E0228 11:52:05.876227 4972 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfa7738-9ed8-475f-863e-ea582e7c1b8b" containerName="oc" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.876247 4972 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfa7738-9ed8-475f-863e-ea582e7c1b8b" containerName="oc" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.876412 4972 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfa7738-9ed8-475f-863e-ea582e7c1b8b" containerName="oc" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.877660 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.898347 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.963425 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7kzv\" (UniqueName: \"kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.963521 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:05 crc kubenswrapper[4972]: I0228 11:52:05.963549 4972 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.066014 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7kzv\" (UniqueName: \"kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.066121 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.066154 4972 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.066739 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.066856 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.085305 4972 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7kzv\" (UniqueName: \"kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv\") pod \"certified-operators-dj2tw\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.220068 4972 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.734516 4972 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:06 crc kubenswrapper[4972]: I0228 11:52:06.863216 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerStarted","Data":"fb76e3dd7170166d551c05c4b53ff9c79aacd6f75e132b9147c538aa6cf2bfe6"} Feb 28 11:52:07 crc kubenswrapper[4972]: I0228 11:52:07.895756 4972 generic.go:334] "Generic (PLEG): container finished" podID="445f2340-8c53-42be-bdf3-b8d330f2e0a3" containerID="3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b" exitCode=0 Feb 28 11:52:07 crc kubenswrapper[4972]: I0228 11:52:07.895867 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerDied","Data":"3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b"} Feb 28 11:52:08 crc kubenswrapper[4972]: I0228 11:52:08.908663 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerStarted","Data":"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828"} Feb 28 11:52:09 crc kubenswrapper[4972]: I0228 11:52:09.925780 4972 generic.go:334] "Generic (PLEG): container finished" podID="445f2340-8c53-42be-bdf3-b8d330f2e0a3" containerID="3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828" exitCode=0 Feb 28 11:52:09 crc kubenswrapper[4972]: I0228 11:52:09.925842 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerDied","Data":"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828"} Feb 28 11:52:11 crc kubenswrapper[4972]: I0228 11:52:11.950805 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerStarted","Data":"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d"} Feb 28 11:52:11 crc kubenswrapper[4972]: I0228 11:52:11.980603 4972 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dj2tw" podStartSLOduration=4.561529229 podStartE2EDuration="6.980573948s" podCreationTimestamp="2026-02-28 11:52:05 +0000 UTC" firstStartedPulling="2026-02-28 11:52:07.900105172 +0000 UTC m=+4584.812112910" lastFinishedPulling="2026-02-28 11:52:10.319149851 +0000 UTC m=+4587.231157629" observedRunningTime="2026-02-28 11:52:11.971819143 +0000 UTC m=+4588.883826901" watchObservedRunningTime="2026-02-28 11:52:11.980573948 +0000 UTC m=+4588.892581726" Feb 28 11:52:16 crc kubenswrapper[4972]: I0228 11:52:16.222625 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:16 crc kubenswrapper[4972]: I0228 11:52:16.223733 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:16 crc kubenswrapper[4972]: I0228 11:52:16.348238 4972 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:17 crc kubenswrapper[4972]: I0228 11:52:17.000876 4972 scope.go:117] "RemoveContainer" containerID="49befc062660d9dd2995c4ee2f3fe57485d96db7627b49121861d1460c2dfad9" Feb 28 11:52:17 crc kubenswrapper[4972]: I0228 11:52:17.080674 4972 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:17 crc kubenswrapper[4972]: I0228 11:52:17.127238 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:19 crc kubenswrapper[4972]: I0228 11:52:19.052074 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dj2tw" podUID="445f2340-8c53-42be-bdf3-b8d330f2e0a3" containerName="registry-server" containerID="cri-o://a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d" gracePeriod=2 Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.030382 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.073807 4972 generic.go:334] "Generic (PLEG): container finished" podID="445f2340-8c53-42be-bdf3-b8d330f2e0a3" containerID="a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d" exitCode=0 Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.073847 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerDied","Data":"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d"} Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.073873 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dj2tw" event={"ID":"445f2340-8c53-42be-bdf3-b8d330f2e0a3","Type":"ContainerDied","Data":"fb76e3dd7170166d551c05c4b53ff9c79aacd6f75e132b9147c538aa6cf2bfe6"} Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.073891 4972 scope.go:117] "RemoveContainer" containerID="a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.073906 4972 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dj2tw" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.103375 4972 scope.go:117] "RemoveContainer" containerID="3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.125740 4972 scope.go:117] "RemoveContainer" containerID="3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173428 4972 scope.go:117] "RemoveContainer" containerID="a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173584 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7kzv\" (UniqueName: \"kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv\") pod \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173747 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities\") pod \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173835 4972 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content\") pod \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\" (UID: \"445f2340-8c53-42be-bdf3-b8d330f2e0a3\") " Feb 28 11:52:20 crc kubenswrapper[4972]: E0228 11:52:20.173878 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d\": container with ID starting with a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d not found: ID does not exist" containerID="a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173929 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d"} err="failed to get container status \"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d\": rpc error: code = NotFound desc = could not find container \"a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d\": container with ID starting with a6ce89f85bd1750894151f84caf8616793c725c981313c8e291dabe2f638152d not found: ID does not exist" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.173964 4972 scope.go:117] "RemoveContainer" containerID="3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.174764 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities" (OuterVolumeSpecName: "utilities") pod "445f2340-8c53-42be-bdf3-b8d330f2e0a3" (UID: "445f2340-8c53-42be-bdf3-b8d330f2e0a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.181113 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv" (OuterVolumeSpecName: "kube-api-access-l7kzv") pod "445f2340-8c53-42be-bdf3-b8d330f2e0a3" (UID: "445f2340-8c53-42be-bdf3-b8d330f2e0a3"). InnerVolumeSpecName "kube-api-access-l7kzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 28 11:52:20 crc kubenswrapper[4972]: E0228 11:52:20.182599 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828\": container with ID starting with 3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828 not found: ID does not exist" containerID="3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.182658 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828"} err="failed to get container status \"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828\": rpc error: code = NotFound desc = could not find container \"3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828\": container with ID starting with 3b65644bf3c7742dd3ee9ab3bd401388f738b24e899fd4bdb7aa9505ad475828 not found: ID does not exist" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.182695 4972 scope.go:117] "RemoveContainer" containerID="3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b" Feb 28 11:52:20 crc kubenswrapper[4972]: E0228 11:52:20.183040 4972 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b\": container with ID starting with 3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b not found: ID does not exist" containerID="3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.183076 4972 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b"} err="failed to get container status \"3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b\": rpc error: code = NotFound desc = could not find container \"3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b\": container with ID starting with 3c53156b0a726800dd39f3e96fd67de16e6455c6042ee2c7e37bf2b398d5f46b not found: ID does not exist" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.228852 4972 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "445f2340-8c53-42be-bdf3-b8d330f2e0a3" (UID: "445f2340-8c53-42be-bdf3-b8d330f2e0a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.276749 4972 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.276801 4972 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7kzv\" (UniqueName: \"kubernetes.io/projected/445f2340-8c53-42be-bdf3-b8d330f2e0a3-kube-api-access-l7kzv\") on node \"crc\" DevicePath \"\"" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.276816 4972 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445f2340-8c53-42be-bdf3-b8d330f2e0a3-utilities\") on node \"crc\" DevicePath \"\"" Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.412322 4972 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:20 crc kubenswrapper[4972]: I0228 11:52:20.426724 4972 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dj2tw"] Feb 28 11:52:21 crc kubenswrapper[4972]: I0228 11:52:21.801198 4972 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445f2340-8c53-42be-bdf3-b8d330f2e0a3" path="/var/lib/kubelet/pods/445f2340-8c53-42be-bdf3-b8d330f2e0a3/volumes" Feb 28 11:52:28 crc kubenswrapper[4972]: I0228 11:52:28.891136 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:52:28 crc kubenswrapper[4972]: I0228 11:52:28.891847 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:52:58 crc kubenswrapper[4972]: I0228 11:52:58.890986 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:52:58 crc kubenswrapper[4972]: I0228 11:52:58.891839 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:53:28 crc kubenswrapper[4972]: I0228 11:53:28.891128 4972 patch_prober.go:28] interesting pod/machine-config-daemon-w2q4v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 28 11:53:28 crc kubenswrapper[4972]: I0228 11:53:28.892054 4972 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 28 11:53:28 crc kubenswrapper[4972]: I0228 11:53:28.892140 4972 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" Feb 28 11:53:28 crc kubenswrapper[4972]: I0228 11:53:28.893153 4972 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893"} pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 28 11:53:28 crc kubenswrapper[4972]: I0228 11:53:28.893253 4972 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" containerName="machine-config-daemon" containerID="cri-o://3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893" gracePeriod=600 Feb 28 11:53:29 crc kubenswrapper[4972]: E0228 11:53:29.025999 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:53:29 crc kubenswrapper[4972]: I0228 11:53:29.981062 4972 generic.go:334] "Generic (PLEG): container finished" podID="118d9c89-cc02-47b8-886e-e72848028ff2" containerID="3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893" exitCode=0 Feb 28 11:53:29 crc kubenswrapper[4972]: I0228 11:53:29.981161 4972 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" event={"ID":"118d9c89-cc02-47b8-886e-e72848028ff2","Type":"ContainerDied","Data":"3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893"} Feb 28 11:53:29 crc kubenswrapper[4972]: I0228 11:53:29.981599 4972 scope.go:117] "RemoveContainer" containerID="76d61b01d9425461bcb524f5f6e216e0046315ca56916a1740b7addfabdb02e5" Feb 28 11:53:29 crc kubenswrapper[4972]: I0228 11:53:29.983898 4972 scope.go:117] "RemoveContainer" containerID="3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893" Feb 28 11:53:29 crc kubenswrapper[4972]: E0228 11:53:29.984592 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" Feb 28 11:53:44 crc kubenswrapper[4972]: I0228 11:53:44.789645 4972 scope.go:117] "RemoveContainer" containerID="3ae9e1878c51d1b249626d40e82122b4c801f49155e6f7ac7028a7213f7a0893" Feb 28 11:53:44 crc kubenswrapper[4972]: E0228 11:53:44.790672 4972 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w2q4v_openshift-machine-config-operator(118d9c89-cc02-47b8-886e-e72848028ff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-w2q4v" podUID="118d9c89-cc02-47b8-886e-e72848028ff2" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515150553531024450 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015150553531017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015150542073016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015150542073015457 5ustar corecore