[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [PATCH 4/6] xen: remove xen dependency in drivers, ether, mempool



Signed-off-by: Jianfeng Tan <jianfeng.tan@xxxxxxxxx>
---
 drivers/crypto/qat/qat_qp.c      | 7 +------
 drivers/net/i40e/i40e_rxtx.c     | 8 ++------
 lib/librte_ether/rte_ethdev.c    | 7 +------
 lib/librte_mempool/rte_mempool.c | 8 ++------
 4 files changed, 6 insertions(+), 24 deletions(-)

diff --git a/drivers/crypto/qat/qat_qp.c b/drivers/crypto/qat/qat_qp.c
index 5048d21..34f75ca 100644
--- a/drivers/crypto/qat/qat_qp.c
+++ b/drivers/crypto/qat/qat_qp.c
@@ -122,14 +122,9 @@ queue_dma_zone_reserve(const char *queue_name, uint32_t 
queue_size,
        break;
        default:
                memzone_flags = RTE_MEMZONE_SIZE_HINT_ONLY;
-}
-#ifdef RTE_LIBRTE_XEN_DOM0
-       return rte_memzone_reserve_bounded(queue_name, queue_size,
-               socket_id, 0, RTE_CACHE_LINE_SIZE, RTE_PGSIZE_2M);
-#else
+       }
        return rte_memzone_reserve_aligned(queue_name, queue_size, socket_id,
                memzone_flags, queue_size);
-#endif
 }
 
 int qat_crypto_sym_qp_setup(struct rte_cryptodev *dev, uint16_t queue_pair_id,
diff --git a/drivers/net/i40e/i40e_rxtx.c b/drivers/net/i40e/i40e_rxtx.c
index d42c23c..f571e79 100644
--- a/drivers/net/i40e/i40e_rxtx.c
+++ b/drivers/net/i40e/i40e_rxtx.c
@@ -2221,12 +2221,8 @@ i40e_memzone_reserve(const char *name, uint32_t len, int 
socket_id)
        if (mz)
                return mz;
 
-       if (rte_xen_dom0_supported())
-               mz = rte_memzone_reserve_bounded(name, len,
-                               socket_id, 0, I40E_RING_BASE_ALIGN, 
RTE_PGSIZE_2M);
-       else
-               mz = rte_memzone_reserve_aligned(name, len,
-                               socket_id, 0, I40E_RING_BASE_ALIGN);
+       mz = rte_memzone_reserve_aligned(name, len,
+                                        socket_id, 0, I40E_RING_BASE_ALIGN);
        return mz;
 }
 
diff --git a/lib/librte_ether/rte_ethdev.c b/lib/librte_ether/rte_ethdev.c
index 0597641..cb0bde7 100644
--- a/lib/librte_ether/rte_ethdev.c
+++ b/lib/librte_ether/rte_ethdev.c
@@ -2818,12 +2818,7 @@ rte_eth_dma_zone_reserve(const struct rte_eth_dev *dev, 
const char *ring_name,
        if (mz)
                return mz;
 
-       if (rte_xen_dom0_supported())
-               return rte_memzone_reserve_bounded(z_name, size, socket_id,
-                                                  0, align, RTE_PGSIZE_2M);
-       else
-               return rte_memzone_reserve_aligned(z_name, size, socket_id,
-                                                  0, align);
+       return rte_memzone_reserve_aligned(z_name, size, socket_id, 0, align);
 }
 
 int
diff --git a/lib/librte_mempool/rte_mempool.c b/lib/librte_mempool/rte_mempool.c
index 6fc3c9c..6d726ae 100644
--- a/lib/librte_mempool/rte_mempool.c
+++ b/lib/librte_mempool/rte_mempool.c
@@ -527,11 +527,7 @@ rte_mempool_populate_default(struct rte_mempool *mp)
        if (mp->nb_mem_chunks != 0)
                return -EEXIST;
 
-       if (rte_xen_dom0_supported()) {
-               pg_sz = RTE_PGSIZE_2M;
-               pg_shift = rte_bsf32(pg_sz);
-               align = pg_sz;
-       } else if (rte_eal_has_hugepages()) {
+       if (rte_eal_has_hugepages()) {
                pg_shift = 0; /* not needed, zone is physically contiguous */
                pg_sz = 0;
                align = RTE_CACHE_LINE_SIZE;
@@ -568,7 +564,7 @@ rte_mempool_populate_default(struct rte_mempool *mp)
                else
                        paddr = mz->phys_addr;
 
-               if (rte_eal_has_hugepages() && !rte_xen_dom0_supported())
+               if (rte_eal_has_hugepages())
                        ret = rte_mempool_populate_phys(mp, mz->addr,
                                paddr, mz->len,
                                rte_mempool_memchunk_mz_free,
-- 
2.7.4


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.