[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Xen-devel] [PATCH v5 00/16] x86/hvm: I/O emulation cleanup and fix
Il 30/06/2015 16:48, Fabio Fantoni ha
scritto:
Il
30/06/2015 15:05, Paul Durrant ha scritto:
This patch series re-works much of the
code involved in emulation of port
and memory mapped I/O for HVM guests.
The code has become very convoluted and, at least by inspection,
certain
emulations will apparently malfunction.
The series is broken down into 16 patches (which are also
available in
my xenbits repo:
http://xenbits.xen.org/gitweb/?p=people/pauldu/xen.git
on the emulation32 branch).
Previous changelog
------------------
v2:
- Removed bogus assertion from patch #15
- Re-worked patch #17 after basic testing of back-port onto
XenServer
v3:
- Addressed comments from Jan
- Re-ordered series to bring a couple of more trivial patches
to the
front
- Backport to XenServer (4.5) now passing automated tests
- Tested on unstable with QEMU upstream and trad, with and
without
HAP (to force shadow emulation)
v4:
- Removed previous patch (make sure translated MMIO reads or
writes fall within a page) and rebased rest of series.
- Address Jan's comments on patch #1
Changelog (now per-patch)
-------------------------
0001-x86-hvm-make-sure-emulation-is-retried-if-domain-is-.patch
This is a fix for an issue on staging reported by Don Slutz
0002-x86-hvm-remove-multiple-open-coded-chunking-loops.patch
v5: Addressed further comments from Jan
0003-x86-hvm-change-hvm_mmio_read_t-and-hvm_mmio_write_t-.patch
v5: New patch to tidy up types
0004-x86-hvm-restrict-port-numbers-to-uint16_t-and-sizes-.patch
v5: New patch to tidy up more types
0005-x86-hvm-unify-internal-portio-and-mmio-intercepts.patch
v5: Addressed further comments from Jan and simplified
implementation
by passing ioreq_t to accept() function
0006-x86-hvm-add-length-to-mmio-check-op.patch
v5: Simplified by leaving mmio_check() implementation alone and
calling to check last byte if first-byte check passes
0007-x86-hvm-unify-dpci-portio-intercept-with-standard-po.patch
v5: Addressed further comments from Jan
0008-x86-hvm-unify-stdvga-mmio-intercept-with-standard-mm.patch
v5: Fixed semantic problems pointed out by Jan
0009-x86-hvm-limit-reps-to-avoid-the-need-to-handle-retry.patch
v5: Addressed further comments from Jan
0010-x86-hvm-only-call-hvm_io_assist-from-hvm_wait_for_io.patch
v5: Added Jan's acked-by
0011-x86-hvm-split-I-O-completion-handling-from-state-mod.patch
v5: Confirmed call to msix_write_completion() is in the correct
place.
0012-x86-hvm-remove-HVMIO_dispatched-I-O-state.patch
v5: Added some extra comments to the commit
0013-x86-hvm-remove-hvm_io_state-enumeration.patch
v5: Added Jan's acked-by
0014-x86-hvm-use-ioreq_t-to-track-in-flight-state.patch
v5: Added missing hunk with call to handle_pio()
0015-x86-hvm-always-re-emulate-I-O-from-a-buffer.patch
v5: Added Jan's acked-by
0016-x86-hvm-track-large-memory-mapped-accesses-by-buffer.patch
v5: Fixed to cache up three distict I/O emulations per
instruction
Testing
-------
The series was been back-ported to staging-4.5 and then dropped
onto the
XenServer (Dundee) patch queue. All automated branch-safety
tests pass.
The series as-is has been manually tested with a Windows 7
(32-bit) VM
using upstream QEMU.
Thanks for your work.
I did some very fast tests, no regression found, on my linux domUs
qxl is still not working and I'm unable to debug it.
@Jim Fehlig: can you try if qxl is still working at least on suse
dom0/domU after this serie please?
Can someone tell me how to debug the qxl problem now that qemu
don't crash anymore but remain at 100% cpu without nothing about
in dom0 logs (if exists) please?
Thanks for any reply and sorry for my bad english.
I don't have knowledge about x86
emulation but I'm trying desperately
to find the cause of such problems
which persists despite this serie of patches.
In latest xengt xen patches I saw this patch: "vgt:
add support of emulating SSE2 instruction MOVD"
https://github.com/01org/XenGT-Preview-xen/commit/f2bad31f80f698a452c37cb39841da8e4f69350f
This xengt patch is still based on xen 4.5 instead, and
x86_emulate.c is different but I noticed some strange things looking
at it and also comparing it with upstream staging...
xengt add this:
case 0x7e: /* movd
xmm,mm/mm32 */
this seems still missing in upstream staging, should it be added?
Should I try the patch with latest upstream xen or does it need
other changes?
There is also this:
ea.bytes = (b ==
0x7e ? 4
: 16);
An sse2 istruction of 4 byte seems strange to me, is it right?
In upstream I saw this:
case 0xe7: /* movntq mm,m64 */
/* {,v}movntdq xmm,m128 */
/* vmovntdq ymm,m256 */
fail_if(ea.type != OP_MEM);
fail_if(vex.pfx == vex_f3);
/* fall through */
In the last qemu crash I had with qxl on linux I got this:
#0 __memset_sse2 () at ../sysdeps/x86_64/multiarch/../memset.S:908
Latest istruction:
=> 0x7ffff3713f7b <__memset_sse2+2363>: movntdq %xmm0,(%rdi)
Is it possible that the "fail_if(ea.type != OP_MEM);" or the other
one make a difference?
After applying this patch serie qemu doesn't crash anymore but it
remains at 100% cpu and is unusable, I can do only xl destroy, I
find nothing in logs and I'm unable to debug it.
Is there something I can do to debug this?
I also took a fast look at suse kernel patches
(https://github.com/openSUSE/kernel-source/tree/SLE12-SP1) where qxl
is also working on linux domUs (other things seems already similar
based on what Jim Fehlig told me) but I didn't find a possible
fix/workaround for it to try. Can someone tell me about possible
patches I should try please?
Any help to find the workaround or fix to apply upstream and have
linux domUs working with qxl in every case is appreciated.
|
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|