[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [v8][PATCH 06/16] hvmloader/pci: disable all pci devices conflicting with rdm



On Thu, Jul 16, 2015 at 5:18 PM, George Dunlap
<George.Dunlap@xxxxxxxxxxxxx> wrote:
> On Thu, Jul 16, 2015 at 4:39 PM, George Dunlap <george.dunlap@xxxxxxxxxx> 
> wrote:
>> On 07/16/2015 04:20 PM, Chen, Tiejun wrote:
>>>>> What about this?
>>>>
>>>> Looks reasonable (but don't forget that I continue to be unconvinced
>>>> that the patch as a whole makes sense).
>>>
>>> Yes, I always keep this in my mind as I mentioned in patch #00. Any risk
>>> you're still concerning? Is it that case if guest OS force enabling
>>> these devices again? IMO, at this point there are two cases:
>>>
>>> #1. Without passing through a RMRR device
>>>
>>> Those emulated devices don't create 1:1 mapping so its safe, right?
>>>
>>> #2. With passing through a RMRR device
>>>
>>> This just probably cause these associated devices not to work well, but
>>> still don't bring any impact to other Domains, right? I mean this isn't
>>> going to worsen the preexisting situation.
>>>
>>> If I'm wrong please correct me.
>>
>> But I think the issue is, without doing *something* about MMIO
>> collisions, the feature as a whole is sort of pointless.  You can
>> carefully specify rdm="strategy=host,reserved=strict", but you might
>> still get devices whose MMIO regions conflict with RMMRs, and there's
>> nothing you can really do about it.
>>
>> And although I personally think it might be possible / reasonable to
>> check in a newly-written, partial MMIO collision avoidance patch, not
>> everyone might agree.  Even if I were to rewrite and post a patch
>> myself, they may argue that doing such a complicated re-design after the
>> feature freeze shouldn't be allowed.
>
> What about something like this?
>
>  -George
>
> ---
>  [PATCH] hvmloader/pci: Try to avoid placing BARs in RMRRs
>
> Try to avoid placing PCI BARs over RMRRs:
>
> - If mmio_hole_size is not specified, and the existing MMIO range has
>   RMRRs in it, and there is space to expand the hole in lowmem without
>   moving more memory, then make the MMIO hole as large as possible.
>
> - When placing RMRRs, find the next RMRR higher than the current base
>   in the lowmem mmio hole.  If it overlaps, skip ahead of it and find
>   the next one.
>
> This certainly won't work in all cases, but it should work in a
> significant number of cases.  Additionally, users should be able to
> work around problems by setting mmio_hole_size larger in the guest
> config.
>
> Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
> ---
> THIS WILL NOT COMPILE, as it needs check_overlap_all() to be implemented.
>
> It's just a proof-of-concept for discussion.
> ---
>  tools/firmware/hvmloader/pci.c | 42 
> ++++++++++++++++++++++++++++++++++++++++++
>  1 file changed, 42 insertions(+)
>
> diff --git a/tools/firmware/hvmloader/pci.c b/tools/firmware/hvmloader/pci.c
> index 5ff87a7..dcb8cd0 100644
> --- a/tools/firmware/hvmloader/pci.c
> +++ b/tools/firmware/hvmloader/pci.c
> @@ -38,6 +38,25 @@ uint64_t pci_hi_mem_start = 0, pci_hi_mem_end = 0;
>  enum virtual_vga virtual_vga = VGA_none;
>  unsigned long igd_opregion_pgbase = 0;
>
> +/* Find the lowest RMRR higher than base */
> +int find_next_rmrr(uint32_t base)
> +{
> +    int next_rmrr=-1;
> +    uing64_t min_base = (1ull << 32);
> +
> +    for ( i = 0; i < memory_map.nr_map ; i++ )
> +    {
> +        if ( memory_map.map[i].type == E820_RESERVED
> +             && memory_map.map[i].addr > base
> +             && memory_map.map[i].addr < min_base)
> +        {
> +            next_rmrr = i;
> +            min_base = memory_map.map[i].addr;
> +        }
> +    }
> +    return next_rmrr;
> +}
> +
>  void pci_setup(void)
>  {
>      uint8_t is_64bar, using_64bar, bar64_relocate = 0;
> @@ -299,6 +318,15 @@ void pci_setup(void)
>                      || (((pci_mem_start << 1) >> PAGE_SHIFT)
>                          >= hvm_info->low_mem_pgend)) )
>              pci_mem_start <<= 1;
> +
> +        /*
> +         * Try to accomodate RMRRs in our MMIO region on a best-effort basis.
> +         * If we have RMRRs in the range, then make pci_mem_start just after
> +         * hvm_info->low_mem_pgend.
> +         */
> +        if ( pci_mem_start > (hvm_info->low_mem_pgend << PAGE_SHIFT) &&
> +             check_overlap_all(pci_mem_start, pci_mem_end-pci_mem_start) )
> +            pci_mem_start = (hvm_info->low_mem_pgend + 1 ) << PAGE_SHIFT);
>      }
>
>      if ( mmio_total > (pci_mem_end - pci_mem_start) )
> @@ -352,6 +380,8 @@ void pci_setup(void)
>      io_resource.base = 0xc000;
>      io_resource.max = 0x10000;
>
> +    next_rmrr = find_next_rmrr(pci_mem_start);
> +
>      /* Assign iomem and ioport resources in descending order of size. */
>      for ( i = 0; i < nr_bars; i++ )
>      {
> @@ -407,6 +437,18 @@ void pci_setup(void)
>          }
>
>          base = (resource->base  + bar_sz - 1) & ~(uint64_t)(bar_sz - 1);
> +
> +        /* If we're using mem_resource, check for RMRR conflicts */
> +        while ( resource == &mem_resource &&
> +                next_rmrr > 0 &&
> +                check_overlap(base, bar_sz,
> +                              memory_map.map[next_rmrr].addr,
> +                              memory_map.map[next_rmrr].size)) {
> +            base = memory_map.map[next_rmrr].addr +
> memory_map.map[next_rmrr].size;
> +            base = (resource->base  + bar_sz - 1) & ~(uint64_t)(bar_sz - 1);

Sorry, this should obviously be

base = (base  + bar_sz - 1) & ~(uint64_t)(bar_sz - 1);

I thought I'd changed it, but apparently I just skipped that step. :-)

 -George

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.