|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH v1 12/14] arm/libxl: Emulated PCI device tree node in libxl
Hi Julien,
> On 9 Sep 2021, at 2:59 pm, Julien Grall <julien@xxxxxxx> wrote:
>
>
>
> On 20/08/2021 17:03, Rahul Singh wrote:
>> Hi Julien,
>
> Hi Rahul,
>
>>> On 19 Aug 2021, at 2:00 pm, Julien Grall <julien@xxxxxxx> wrote:
>>>
>>> Hi Rahul,
>>>
>>> On 19/08/2021 13:02, Rahul Singh wrote:
>>>> libxl will create an emulated PCI device tree node in the device tree to
>>>> enable the guest OS to discover the virtual PCI during guest boot.
>>>> Emulated PCI device tree node will only be created when there is any
>>>> device assigned to guest.
>>>> A new area has been reserved in the arm guest physical map at
>>>> which the VPCI bus is declared in the device tree (reg and ranges
>>>> parameters of the node).
>>>> Signed-off-by: Rahul Singh <rahul.singh@xxxxxxx>
>>>> ---
>>>> tools/libs/light/libxl_arm.c | 109 ++++++++++++++++++++++++++
>>>> tools/libs/light/libxl_types.idl | 1 +
>>>> tools/xl/xl_parse.c | 2 +
>>>> xen/include/public/arch-arm.h | 11 +++
>>>> xen/include/public/device_tree_defs.h | 1 +
>>>> 5 files changed, 124 insertions(+)
>>>> diff --git a/tools/libs/light/libxl_arm.c b/tools/libs/light/libxl_arm.c
>>>> index e3140a6e00..a091e97e76 100644
>>>> --- a/tools/libs/light/libxl_arm.c
>>>> +++ b/tools/libs/light/libxl_arm.c
>>>> @@ -269,6 +269,58 @@ static int fdt_property_regs(libxl__gc *gc, void *fdt,
>>>> return fdt_property(fdt, "reg", regs, sizeof(regs));
>>>> }
>>>> +static int fdt_property_values(libxl__gc *gc, void *fdt,
>>>> + const char *name, unsigned num_cells, ...)
>>>> +{
>>>> + uint32_t prop[num_cells];
>>>> + be32 *cells = &prop[0];
>>>> + int i;
>>>> + va_list ap;
>>>> + uint32_t arg;
>>>> +
>>>> + va_start(ap, num_cells);
>>>> + for (i = 0 ; i < num_cells; i++) {
>>>> + arg = va_arg(ap, uint32_t);
>>>> + set_cell(&cells, 1, arg);
>>>> + }
>>>> + va_end(ap);
>>>> +
>>>> + return fdt_property(fdt, name, prop, sizeof(prop));
>>>> +}
>>>> +
>>>> +static int fdt_property_vpci_ranges(libxl__gc *gc, void *fdt,
>>>> + unsigned addr_cells,
>>>> + unsigned size_cells,
>>>> + unsigned num_regs, ...)
>>>> +{
>>>> + uint32_t regs[num_regs*((addr_cells*2)+size_cells+1)];
>>>> + be32 *cells = ®s[0];
>>>> + int i;
>>>> + va_list ap;
>>>> + uint64_t arg;
>>>> +
>>>> + va_start(ap, num_regs);
>>>> + for (i = 0 ; i < num_regs; i++) {
>>>> + /* Set the memory bit field */
>>>> + arg = va_arg(ap, uint64_t);
>>>> + set_cell(&cells, 1, arg);
>>>> +
>>>> + /* Set the vpci bus address */
>>>> + arg = addr_cells ? va_arg(ap, uint64_t) : 0;
>>>> + set_cell(&cells, addr_cells , arg);
>>>> +
>>>> + /* Set the cpu bus address where vpci address is mapped */
>>>> + set_cell(&cells, addr_cells, arg);
>>>> +
>>>> + /* Set the vpci size requested */
>>>> + arg = size_cells ? va_arg(ap, uint64_t) : 0;
>>>> + set_cell(&cells, size_cells,arg);
>>>> + }
>>>> + va_end(ap);
>>>> +
>>>> + return fdt_property(fdt, "ranges", regs, sizeof(regs));
>>>> +}
>>>> +
>>>> static int make_root_properties(libxl__gc *gc,
>>>> const libxl_version_info *vers,
>>>> void *fdt)
>>>> @@ -668,6 +720,57 @@ static int make_vpl011_uart_node(libxl__gc *gc, void
>>>> *fdt,
>>>> return 0;
>>>> }
>>>> +static int make_vpci_node(libxl__gc *gc, void *fdt,
>>>> + const struct arch_info *ainfo,
>>>> + struct xc_dom_image *dom)
>>>> +{
>>>> + int res;
>>>> + const uint64_t vpci_ecam_base = GUEST_VPCI_ECAM_BASE;
>>>> + const uint64_t vpci_ecam_size = GUEST_VPCI_ECAM_SIZE;
>>>> + const char *name = GCSPRINTF("pcie@%"PRIx64, vpci_ecam_base);
>>>> +
>>>> + res = fdt_begin_node(fdt, name);
>>>> + if (res) return res;
>>>> +
>>>> + res = fdt_property_compat(gc, fdt, 1, "pci-host-ecam-generic");
>>>> + if (res) return res;
>>>> +
>>>> + res = fdt_property_string(fdt, "device_type", "pci");
>>>> + if (res) return res;
>>>> +
>>>> + res = fdt_property_regs(gc, fdt, GUEST_ROOT_ADDRESS_CELLS,
>>>> + GUEST_ROOT_SIZE_CELLS, 1, vpci_ecam_base, vpci_ecam_size);
>>>> + if (res) return res;
>>>> +
>>>> + res = fdt_property_values(gc, fdt, "bus-range", 2, 0,17);
>>>
>>> AFAICT, the "bus-range" is optional. Can you explain why we need it?
>> We need it to implement the function pci_ecam_map_bus().
>
> Ok. Then why next question is what does the 17 mean? Is it tie to how we
> implement the vPCI in Xen or the region we reserved?
>
Ah. I forgot to change it to 0xff. It should be 0xff.
> [...]
>
>>>
>>>> +
>>>> if (b_info->type != LIBXL_DOMAIN_TYPE_PV)
>>>> return;
>>>> diff --git a/tools/libs/light/libxl_types.idl
>>>> b/tools/libs/light/libxl_types.idl
>>>> index 3f9fff653a..78b1ddf0b8 100644
>>>> --- a/tools/libs/light/libxl_types.idl
>>>> +++ b/tools/libs/light/libxl_types.idl
>>>> @@ -644,6 +644,7 @@ libxl_domain_build_info = Struct("domain_build_info",[
>>>> ("arch_arm", Struct(None, [("gic_version", libxl_gic_version),
>>>> ("vuart", libxl_vuart_type),
>>>> + ("vpci", libxl_defbool),
>>>
>>> Any new addition in the structure should be accompanied with a LIBXL_HAVE_*
>>> in the libxl.h header.
>> OK.
>>>
>>>> ])),
>>>> ("arch_x86", Struct(None, [("msr_relaxed", libxl_defbool),
>>>> ])),
>>>> diff --git a/tools/xl/xl_parse.c b/tools/xl/xl_parse.c
>>>> index 17dddb4cd5..ffafbeffb4 100644
>>>> --- a/tools/xl/xl_parse.c
>>>> +++ b/tools/xl/xl_parse.c
>>>> @@ -1497,6 +1497,8 @@ void parse_config_data(const char *config_source,
>>>> }
>>>> if (d_config->num_pcidevs && c_info->type == LIBXL_DOMAIN_TYPE_PV)
>>>> libxl_defbool_set(&b_info->u.pv.e820_host, true);
>>>> + if (d_config->num_pcidevs)
>>>> + libxl_defbool_set(&b_info->arch_arm.vpci, true);
>>>> }
>>>> if (!xlu_cfg_get_list (config, "dtdev", &dtdevs, 0, 0)) {
>>>> diff --git a/xen/include/public/arch-arm.h b/xen/include/public/arch-arm.h
>>>> index 0a9749e768..01d13e669e 100644
>>>> --- a/xen/include/public/arch-arm.h
>>>> +++ b/xen/include/public/arch-arm.h
>>>> @@ -426,6 +426,17 @@ typedef uint64_t xen_callback_t;
>>>> #define GUEST_VPCI_ECAM_BASE xen_mk_ullong(0x10000000)
>>>> #define GUEST_VPCI_ECAM_SIZE xen_mk_ullong(0x10000000)
>>>> +/* PCI-PCIe memory space types */
>>>> +#define GUEST_VPCI_ADDR_TYPE_PREFETCH_MEM xen_mk_ullong(0x42000000)
>>>> +#define GUEST_VPCI_ADDR_TYPE_MEM xen_mk_ullong(0x02000000)
>>>
>>> What the size of those regions?
>> Non Prefetch Memory: Size 64 MB start at 512 MB
>> Prefetch Memory: Size 128 GB start at 36 GB
>>>
>>>> +
>>>> +/* Guest PCI-PCIe memory space where config space and BAR will be
>>>> available.*/
>>>> +#define GUEST_VPCI_PREFETCH_MEM_ADDR xen_mk_ullong(0x900000000)
>>>> +#define GUEST_VPCI_MEM_ADDR xen_mk_ullong(0x20000000)
>>> So far the memory layout defines the address in ascending order. So please
>>> add that after GUEST_RAM_BANK_BASES_*.
>> Ok.
>>>
>>> However, if I am not mistaken that base address you provide will clash with
>>> RAM bank 1. It also seem to be pretty high which means that this will not
>>> work for 32-bit domain or on CPUs that don't allow offer large IPA bits.
>> Yes I also checked that now that it is having clash with RAM bank 1.
>> There is unused space is guest memory that we can use for Non Prefetch
>> Memory as per below guest memory map.
>> https://gitlab.com/xen-project/fusa/fusa-docs/-/blob/master/high-level/guest-memory-layout-arm.rst
>> Proposed value:
>> Non Prefetch Memory: Size 64 MB start at 0x22001000
>> Prefetch Memory: Size 4 GB start at 4 GB.
>
> The base address looks fine to me. However, the sizes are much smaller to
> what you initially suggested. Would you be able to clarify why the smaller
> sizes are fine?
As per current design we can have 32 device in guest therefore I thought 64 MB
is sufficient ( 2 MB per device ).
>
>>>
>>> I think we need to start making the guest layout more dynamic. The VPCI
>>> memory space would have to go right after the end of the RAM allocated for
>>> a given guest.
>>>
>>>> +
>>>> +#define GUEST_VPCI_PREFETCH_MEM_SIZE xen_mk_ullong(0x2000000000)
>>>> +#define GUEST_VPCI_MEM_SIZE xen_mk_ullong(0x04000000)
>>>
>>> It would be better if the size for each region is defined right after each
>>> base.
>> OK.
>>>
>>> Also, how did you decide the size of each region?
>> I thought 64 MB will be sufficient. I think it should be based on number of
>> devices we can assign to the guest.
>
> We don't have to get the size right now. What I am more interested is to have
> a trace about how those values were decided (even if it just saying random).
> This will help to make any decision if in the future we need to resize (in
> particular downsize) the regions.
As per Stefano suggestion in another mail he suggested 256 MB. I also think 256
MB will be sufficient ( 8 MB per device) . I will comment this in code.
Regards,
Rahul
>
> Cheers,
>
> --
> Julien Grall
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |