[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] RE: [Xen-ia64-devel] [PATCH] fully virtualize psr and ipsr onnon-VTIdomain
Hi Shuji, Welcome to community! Not quite understand the issue. Is the panic related to gcc segmentation fault? Has you applied vcpu_translate patch? Thanks, -Anthony >-----Original Message----- >From: xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx >[mailto:xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of Shuji >Kurihara >Sent: 2006?3?30? 22:35 >To: Magenheimer, Dan (HP Labs Fort Collins) >Cc: xen-ia64-devel@xxxxxxxxxxxxxxxxxxx >Subject: RE: [Xen-ia64-devel] [PATCH] fully virtualize psr and ipsr >onnon-VTIdomain > > >Hi Dan, > >Some time ago, you reported that the fully virtualize psr and ipsr patch >by Anthony caused Linux compilation to crash the system. >This panic seems to be solved by applying the vcpu_translate_patch. >However, the original gcc segmentation fault problem still occurs on >this system, and it still remains to be solved. Details below. > > psr.patch: > >http://lists.xensource.com/archives/html/xen-ia64-devel/2005-11/msg00312.ht >ml > > vcpu_tranalate.patch: > >http://lists.xensource.com/archives/html/xen-ia64-devel/2006-03/msg00328.ht >ml > >The panic occurs when trying to handle a tlb miss following "itc" >instruction. Below is console output: > >(XEN) vcpu_translate: bad physical address: a00000010000a090 >(XEN) translate_domain_pte: bad mpa=000000010000a090 (> 0000000018000000), >vadr=a00000010000a090,pteval=001000010000a761,itir=0000000000000038 >(XEN) lookup_domain_mpa: bad mpa 000000010000a090 (> 0000000018000000 >(XEN) handle_op: can't handle privop at 0xa00000010000a090 >(op=0x000001a7a7a7a7a7) > slot 0 (type=5), ipsr=0000101208026010 >(XEN) priv_emulate: priv_handle_op fails, isr=0000000000000000 >(XEN) $$$$$ PANIC in domain 1 (k6=f000000007f98000): psr.dt off, >trying to deliver nested dtlb! >(XEN) >(XEN) CPU 0 >(XEN) psr : 0000101208026010 ifs : 800000000000040e ip : [<a00000010000a090>] >(XEN) ip is at ??? >(XEN) unat: 0000000000000000 pfs : c00000000000040e rsc : 000000000000000f >(XEN) rnat: 0000000000000000 bsps: 60000fff7fffc160 pr : 000000000555a261 >(XEN) ldrs: 0000000000700000 ccv : 0010000001c585a1 fpsr: 0009804c8a70033f >(XEN) csd : 0000000000000000 ssd : 0000000000000000 >(XEN) b0 : a00000010000a070 b6 : 20000000001f8780 b7 : 0000000000000000 >(XEN) f6 : 000000000000000000000 f7 : 000000000000000000000 >(XEN) f8 : 000000000000000000000 f9 : 000000000000000000000 >(XEN) f10 : 000000000000000000000 f11 : 000000000000000000000 >(XEN) r1 : 60000000000021f0 r2 : 0000000000000000 r3 : 0000000000000308 >(XEN) r8 : 0000000000000000 r9 : 20000000002c64a0 r10 : 0000000000000000 >(XEN) r11 : c00000000000040e r12 : 60000fffffaa7610 r13 : 20000000002d06a0 >(XEN) r14 : 0000000000000030 r15 : 6000000000100000 r16 : 6000000000100000 >(XEN) r17 : 0000000001bf4200 r18 : 0010000001c585a1 r19 : 0001800000000040 >(XEN) r20 : 000000001613c000 r21 : 0000000000000000 r22 : 5fffff0000000000 >(XEN) r23 : 000000001613c000 r24 : 0000000000000038 r25 : 0010000001c585e1 >(XEN) r26 : 0010000001c585a1 r27 : 0000000000000038 r28 : 0000000000000000 >(XEN) r29 : 4000000000001870 r30 : a00000010000a070 r31 : 000000000555a2a1 >(XEN) vcpu_translate: bad physical address: 60000fff7fffc1d0 >(XEN) translate_domain_pte: bad mpa=00000fff7fffc1d0 (> 0000000018000000), >vadr=60000fff7fffc1d0,pteval=00100fff7fffc761,itir=0000000000000038 >(XEN) lookup_domain_mpa: bad mpa 00000fff7fffc1d0 (> 0000000018000000 >(XEN) r32 : f0000000f0000000 r33 : f0000000f0000000 r34 : f0000000f0000000 >(XEN) r35 : f0000000f0000000 r36 : f0000000f0000000 r37 : f4f4f4f4f4f4f4f4 >(XEN) r38 : f4f4f4f4f4f4f4f4 r39 : f4f4f4f4f4f4f4f4 r40 : f4f4f4f4f4f4f4f4 >(XEN) r41 : f4f4f4f4f4f4f4f4 r42 : f4f4f4f4f4f4f4f4 r43 : f4f4f4f4f4f4f4f4 >(XEN) r44 : f4f4f4f4f4f4f4f4 r45 : f4f4f4f4f4f4f4f4 >(XEN) BUG at domain.c:339 >(XEN) bad hyperprivop; ignored >(XEN) iim=0, iip=f0000000040203d0 >(XEN) bad hyperprivop; ignored >(XEN) iim=0, iip=f0000000040203d0 > >One of the above messages: > > (XEN) vcpu_translate: bad physical address: a00000010000a090 > >The address "a00000010000a090" points to the instruction below. > > a00000010000a090: cb 00 64 00 2e 04 [MMI] (p06) itc.d r25;; > >When the VMM tries to get the opcode to call priv_handle_op(), >it seems to trigger a tlb miss, and causes domU to hang. >It seems from the message that domain is in metaphysical mode >after executing "rsm psr.dt" instruction, and the fault address is >in region 5. > >This situation is similar to the problem vcpu_translate_patch >tries to solve. The patch fixes vcpu_translate() so that >the guest OS does not operate in metaphysical mode in such a case. > >We have run the same test program on Xen 3.0-unstable with CSet#9395 >(which includes vcpu_translate patch) and it run throughout the weekend >without causing any panic. On the other hand, the original Xen without >the patch crashes within 2 hours. However, the original gcc >segmentation faults still occured on the system, so neither CSet#8671 >nor #9395 seems to solve the original problem. > >Thanks, >Shuji > >>Hi Anthony -- >> >>Since things have stabilized, I decided to give this patch >>some testing, primarily to see if it might fix the gcc >>segmentation faults that Fujita and I have been seeing. >>Without this patch, I am able to compile Linux 20 times >>on domU; generally 1 or 2 of the compiles fails because >>of the gcc segfault. With the patch, Xen *crashed* >>on the sixth Linux compile (first try) and halfway >>through the first Linux compile (second try). This >>is on a Tiger4, but I currently am not able to get console >>output so I don't have any information about the crash -- >>other than that the machine didn't reboot. Could you see >>if you could reproduce this? >> >>As an aside, turning off the FAST_BREAK, FAST_ACCESS_REFLECT, >>and FAST_RFI features (which your patch turns off) slowed >>down the benchmark by about 4%. >> >>Thanks, >>Dan >> >>> -----Original Message----- >>> From: Xu, Anthony [mailto:anthony.xu@xxxxxxxxx] >>> Sent: Sunday, November 27, 2005 8:22 PM >>> To: Magenheimer, Dan (HP Labs Fort Collins) >>> Cc: xen-ia64-devel@xxxxxxxxxxxxxxxxxxx >>> Subject: [Xen-ia64-devel] [PATCH] fully virtualize psr and >>> ipsr on non-VTI domain >>> >>> Dan, >>> This patch is intended to fully virtualize psr and ipsr on non-VTI >>> domain. >>> Following things are done in this patch. >>> 1, previously when guest reads psr, it always get psr dt rt >>> it equal to >>> 1. that is because HV doesn't restore these information, >>> metaphysical_mode can't present all these information. I save these >>> information into privregs->vpsr. Thus guest can get correct >>> information >>> about dt, rt and it. >>> 2, when guest reads psr, we should only return low 32bits and >>> 35 and 36 >>> bits, previously return all bits. >>> 3, when guest rsm and ssm psr, HV rsm and ssm some bits of current psr >>> which is used by HV, that is not correct, guest rsm and ssm >>> should only >>> impact guest psr(that is regs->ipsr). >>> 4, mistakenly uses guest DCR, guest DCR should impact guest psr when >>> injecting interruption into guest, but not impact guest ipsr. >>> When injecting interruption into guest,The current implementation is >>> Guest ipsr.be=guest dcr.be >>> Guest ipsr.pp=guest dcr.pp >>> Correct implementation should be, >>> Guest psr.be=guest dcr.be >>> Guest psr.pp=guest dcr.pp. >>> >>> Because of above modifications, I turn off FAST_RFI, FAST_BREAK and >>> FAST_ACCESS_REFLECT. >>> >>> Signed-off-by Anthony Xu < anthony.xu@xxxxxxxxx> >>> >>> One question, why do we need to virtualize guest psr.pp and always set >>> guest psr.pp to 1? >>> >>> Thanks >>> -Anthony > >_______________________________________________ >Xen-ia64-devel mailing list >Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx >http://lists.xensource.com/xen-ia64-devel _______________________________________________ Xen-ia64-devel mailing list Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-ia64-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |