Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit a6970317 authored by Oleg Drokin's avatar Oleg Drokin Committed by Trond Myklebust
Browse files

staging/lustre/o2iblnd: remove references to ib_reg_phsy_mr()



Removed references to ib_reg_phsy_mr() and PMR which was added
to deal with some Chelsio specific scenario, but no longer needed
now.

Reported-by: default avatarStephen Rothwell <sfr@canb.auug.org.au>
Signed-off-by: default avatarAmir Shehata <amir.shehata@intel.com>
Signed-off-by: default avatarOleg Drokin <oleg.drokin@intel.com>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Signed-off-by: default avatarTrond Myklebust <trond.myklebust@primarydata.com>
parent 54204010
Loading
Loading
Loading
Loading
+12 −239
Original line number Original line Diff line number Diff line
@@ -1789,140 +1789,6 @@ struct list_head *kiblnd_pool_alloc_node(kib_poolset_t *ps)
	goto again;
	goto again;
}
}


void kiblnd_pmr_pool_unmap(kib_phys_mr_t *pmr)
{
	kib_pmr_pool_t *ppo = pmr->pmr_pool;
	struct ib_mr *mr = pmr->pmr_mr;

	pmr->pmr_mr = NULL;
	kiblnd_pool_free_node(&ppo->ppo_pool, &pmr->pmr_list);
	if (mr != NULL)
		ib_dereg_mr(mr);
}

int kiblnd_pmr_pool_map(kib_pmr_poolset_t *pps, kib_hca_dev_t *hdev,
		    kib_rdma_desc_t *rd, __u64 *iova, kib_phys_mr_t **pp_pmr)
{
	kib_phys_mr_t *pmr;
	struct list_head *node;
	int rc;
	int i;

	node = kiblnd_pool_alloc_node(&pps->pps_poolset);
	if (node == NULL) {
		CERROR("Failed to allocate PMR descriptor\n");
		return -ENOMEM;
	}

	pmr = container_of(node, kib_phys_mr_t, pmr_list);
	if (pmr->pmr_pool->ppo_hdev != hdev) {
		kiblnd_pool_free_node(&pmr->pmr_pool->ppo_pool, node);
		return -EAGAIN;
	}

	for (i = 0; i < rd->rd_nfrags; i++) {
		pmr->pmr_ipb[i].addr = rd->rd_frags[i].rf_addr;
		pmr->pmr_ipb[i].size = rd->rd_frags[i].rf_nob;
	}

	pmr->pmr_mr = ib_reg_phys_mr(hdev->ibh_pd,
				     pmr->pmr_ipb, rd->rd_nfrags,
				     IB_ACCESS_LOCAL_WRITE |
				     IB_ACCESS_REMOTE_WRITE,
				     iova);
	if (!IS_ERR(pmr->pmr_mr)) {
		pmr->pmr_iova = *iova;
		*pp_pmr = pmr;
		return 0;
	}

	rc = PTR_ERR(pmr->pmr_mr);
	CERROR("Failed ib_reg_phys_mr: %d\n", rc);

	pmr->pmr_mr = NULL;
	kiblnd_pool_free_node(&pmr->pmr_pool->ppo_pool, node);

	return rc;
}

static void kiblnd_destroy_pmr_pool(kib_pool_t *pool)
{
	kib_pmr_pool_t *ppo = container_of(pool, kib_pmr_pool_t, ppo_pool);
	kib_phys_mr_t *pmr;
	kib_phys_mr_t *tmp;

	LASSERT(pool->po_allocated == 0);

	list_for_each_entry_safe(pmr, tmp, &pool->po_free_list, pmr_list) {
		LASSERT(pmr->pmr_mr == NULL);
		list_del(&pmr->pmr_list);

		if (pmr->pmr_ipb != NULL) {
			LIBCFS_FREE(pmr->pmr_ipb,
				    IBLND_MAX_RDMA_FRAGS *
				    sizeof(struct ib_phys_buf));
		}

		LIBCFS_FREE(pmr, sizeof(kib_phys_mr_t));
	}

	kiblnd_fini_pool(pool);
	if (ppo->ppo_hdev != NULL)
		kiblnd_hdev_decref(ppo->ppo_hdev);

	LIBCFS_FREE(ppo, sizeof(kib_pmr_pool_t));
}

static inline int kiblnd_pmr_pool_size(int ncpts)
{
	int size = *kiblnd_tunables.kib_pmr_pool_size / ncpts;

	return max(IBLND_PMR_POOL, size);
}

static int kiblnd_create_pmr_pool(kib_poolset_t *ps, int size,
				  kib_pool_t **pp_po)
{
	struct kib_pmr_pool *ppo;
	struct kib_pool *pool;
	kib_phys_mr_t *pmr;
	int i;

	LIBCFS_CPT_ALLOC(ppo, lnet_cpt_table(),
			 ps->ps_cpt, sizeof(kib_pmr_pool_t));
	if (ppo == NULL) {
		CERROR("Failed to allocate PMR pool\n");
		return -ENOMEM;
	}

	pool = &ppo->ppo_pool;
	kiblnd_init_pool(ps, pool, size);

	for (i = 0; i < size; i++) {
		LIBCFS_CPT_ALLOC(pmr, lnet_cpt_table(),
				 ps->ps_cpt, sizeof(kib_phys_mr_t));
		if (pmr == NULL)
			break;

		pmr->pmr_pool = ppo;
		LIBCFS_CPT_ALLOC(pmr->pmr_ipb, lnet_cpt_table(), ps->ps_cpt,
				 IBLND_MAX_RDMA_FRAGS * sizeof(*pmr->pmr_ipb));
		if (pmr->pmr_ipb == NULL)
			break;

		list_add(&pmr->pmr_list, &pool->po_free_list);
	}

	if (i < size) {
		ps->ps_pool_destroy(pool);
		return -ENOMEM;
	}

	ppo->ppo_hdev = kiblnd_current_hdev(ps->ps_net->ibn_dev);
	*pp_po = pool;
	return 0;
}

static void kiblnd_destroy_tx_pool(kib_pool_t *pool)
static void kiblnd_destroy_tx_pool(kib_pool_t *pool)
{
{
	kib_tx_pool_t *tpo = container_of(pool, kib_tx_pool_t, tpo_pool);
	kib_tx_pool_t *tpo = container_of(pool, kib_tx_pool_t, tpo_pool);
@@ -2078,7 +1944,6 @@ static void kiblnd_net_fini_pools(kib_net_t *net)
	cfs_cpt_for_each(i, lnet_cpt_table()) {
	cfs_cpt_for_each(i, lnet_cpt_table()) {
		kib_tx_poolset_t *tps;
		kib_tx_poolset_t *tps;
		kib_fmr_poolset_t *fps;
		kib_fmr_poolset_t *fps;
		kib_pmr_poolset_t *pps;


		if (net->ibn_tx_ps != NULL) {
		if (net->ibn_tx_ps != NULL) {
			tps = net->ibn_tx_ps[i];
			tps = net->ibn_tx_ps[i];
@@ -2089,11 +1954,6 @@ static void kiblnd_net_fini_pools(kib_net_t *net)
			fps = net->ibn_fmr_ps[i];
			fps = net->ibn_fmr_ps[i];
			kiblnd_fini_fmr_poolset(fps);
			kiblnd_fini_fmr_poolset(fps);
		}
		}

		if (net->ibn_pmr_ps != NULL) {
			pps = net->ibn_pmr_ps[i];
			kiblnd_fini_poolset(&pps->pps_poolset);
		}
	}
	}


	if (net->ibn_tx_ps != NULL) {
	if (net->ibn_tx_ps != NULL) {
@@ -2105,18 +1965,13 @@ static void kiblnd_net_fini_pools(kib_net_t *net)
		cfs_percpt_free(net->ibn_fmr_ps);
		cfs_percpt_free(net->ibn_fmr_ps);
		net->ibn_fmr_ps = NULL;
		net->ibn_fmr_ps = NULL;
	}
	}

	if (net->ibn_pmr_ps != NULL) {
		cfs_percpt_free(net->ibn_pmr_ps);
		net->ibn_pmr_ps = NULL;
	}
}
}


static int kiblnd_net_init_pools(kib_net_t *net, __u32 *cpts, int ncpts)
static int kiblnd_net_init_pools(kib_net_t *net, __u32 *cpts, int ncpts)
{
{
	unsigned long flags;
	unsigned long flags;
	int cpt;
	int cpt;
	int rc;
	int		rc = 0;
	int i;
	int i;


	read_lock_irqsave(&kiblnd_data.kib_global_lock, flags);
	read_lock_irqsave(&kiblnd_data.kib_global_lock, flags);
@@ -2137,12 +1992,16 @@ static int kiblnd_net_init_pools(kib_net_t *net, __u32 *cpts, int ncpts)
		goto failed;
		goto failed;
	}
	}


	/* TX pool must be created later than FMR/PMR, see LU-2268
	/*
	 * for details */
	 * TX pool must be created later than FMR, see LU-2268
	 * for details
	 */
	LASSERT(net->ibn_tx_ps == NULL);
	LASSERT(net->ibn_tx_ps == NULL);


	/* premapping can fail if ibd_nmr > 1, so we always create
	/*
	 * FMR/PMR pool and map-on-demand if premapping failed */
	 * premapping can fail if ibd_nmr > 1, so we always create
	 * FMR pool and map-on-demand if premapping failed
	 */


	net->ibn_fmr_ps = cfs_percpt_alloc(lnet_cpt_table(),
	net->ibn_fmr_ps = cfs_percpt_alloc(lnet_cpt_table(),
					   sizeof(kib_fmr_poolset_t));
					   sizeof(kib_fmr_poolset_t));
@@ -2158,7 +2017,7 @@ static int kiblnd_net_init_pools(kib_net_t *net, __u32 *cpts, int ncpts)
					     kiblnd_fmr_pool_size(ncpts),
					     kiblnd_fmr_pool_size(ncpts),
					     kiblnd_fmr_flush_trigger(ncpts));
					     kiblnd_fmr_flush_trigger(ncpts));
		if (rc == -ENOSYS && i == 0) /* no FMR */
		if (rc == -ENOSYS && i == 0) /* no FMR */
			break; /* create PMR pool */
			break;


		if (rc != 0) { /* a real error */
		if (rc != 0) { /* a real error */
			CERROR("Can't initialize FMR pool for CPT %d: %d\n",
			CERROR("Can't initialize FMR pool for CPT %d: %d\n",
@@ -2175,38 +2034,8 @@ static int kiblnd_net_init_pools(kib_net_t *net, __u32 *cpts, int ncpts)
	cfs_percpt_free(net->ibn_fmr_ps);
	cfs_percpt_free(net->ibn_fmr_ps);
	net->ibn_fmr_ps = NULL;
	net->ibn_fmr_ps = NULL;


	CWARN("Device does not support FMR, failing back to PMR\n");
	CWARN("Device does not support FMR\n");

	if (*kiblnd_tunables.kib_pmr_pool_size <
	    *kiblnd_tunables.kib_ntx / 4) {
		CERROR("Can't set pmr pool size (%d) < ntx / 4(%d)\n",
		       *kiblnd_tunables.kib_pmr_pool_size,
		       *kiblnd_tunables.kib_ntx / 4);
		rc = -EINVAL;
			goto failed;
			goto failed;
	}

	net->ibn_pmr_ps = cfs_percpt_alloc(lnet_cpt_table(),
					   sizeof(kib_pmr_poolset_t));
	if (net->ibn_pmr_ps == NULL) {
		CERROR("Failed to allocate PMR pool array\n");
		rc = -ENOMEM;
		goto failed;
	}

	for (i = 0; i < ncpts; i++) {
		cpt = (cpts == NULL) ? i : cpts[i];
		rc = kiblnd_init_poolset(&net->ibn_pmr_ps[cpt]->pps_poolset,
					 cpt, net, "PMR",
					 kiblnd_pmr_pool_size(ncpts),
					 kiblnd_create_pmr_pool,
					 kiblnd_destroy_pmr_pool, NULL, NULL);
		if (rc != 0) {
			CERROR("Can't initialize PMR pool for CPT %d: %d\n",
			       cpt, rc);
			goto failed;
		}
	}


 create_tx_pool:
 create_tx_pool:
	net->ibn_tx_ps = cfs_percpt_alloc(lnet_cpt_table(),
	net->ibn_tx_ps = cfs_percpt_alloc(lnet_cpt_table(),
@@ -2318,17 +2147,13 @@ void kiblnd_hdev_destroy(kib_hca_dev_t *hdev)
static int kiblnd_hdev_setup_mrs(kib_hca_dev_t *hdev)
static int kiblnd_hdev_setup_mrs(kib_hca_dev_t *hdev)
{
{
	struct ib_mr *mr;
	struct ib_mr *mr;
	int i;
	int rc;
	int rc;
	__u64 mm_size;
	__u64 mr_size;
	int acflags = IB_ACCESS_LOCAL_WRITE | IB_ACCESS_REMOTE_WRITE;
	int acflags = IB_ACCESS_LOCAL_WRITE | IB_ACCESS_REMOTE_WRITE;


	rc = kiblnd_hdev_get_attr(hdev);
	rc = kiblnd_hdev_get_attr(hdev);
	if (rc != 0)
	if (rc != 0)
		return rc;
		return rc;


	if (hdev->ibh_mr_shift == 64) {
		LIBCFS_ALLOC(hdev->ibh_mrs, 1 * sizeof(*hdev->ibh_mrs));
		LIBCFS_ALLOC(hdev->ibh_mrs, 1 * sizeof(*hdev->ibh_mrs));
		if (hdev->ibh_mrs == NULL) {
		if (hdev->ibh_mrs == NULL) {
			CERROR("Failed to allocate MRs table\n");
			CERROR("Failed to allocate MRs table\n");
@@ -2347,53 +2172,6 @@ static int kiblnd_hdev_setup_mrs(kib_hca_dev_t *hdev)


		hdev->ibh_mrs[0] = mr;
		hdev->ibh_mrs[0] = mr;


		goto out;
	}

	mr_size = 1ULL << hdev->ibh_mr_shift;
	mm_size = (unsigned long)high_memory - PAGE_OFFSET;

	hdev->ibh_nmrs = (int)((mm_size + mr_size - 1) >> hdev->ibh_mr_shift);

	if (hdev->ibh_mr_shift < 32 || hdev->ibh_nmrs > 1024) {
		/* it's 4T..., assume we will re-code at that time */
		CERROR("Can't support memory size: x%#llx with MR size: x%#llx\n",
		       mm_size, mr_size);
		return -EINVAL;
	}

	/* create an array of MRs to cover all memory */
	LIBCFS_ALLOC(hdev->ibh_mrs, sizeof(*hdev->ibh_mrs) * hdev->ibh_nmrs);
	if (hdev->ibh_mrs == NULL) {
		CERROR("Failed to allocate MRs' table\n");
		return -ENOMEM;
	}

	for (i = 0; i < hdev->ibh_nmrs; i++) {
		struct ib_phys_buf ipb;
		__u64 iova;

		ipb.size = hdev->ibh_mr_size;
		ipb.addr = i * mr_size;
		iova = ipb.addr;

		mr = ib_reg_phys_mr(hdev->ibh_pd, &ipb, 1, acflags, &iova);
		if (IS_ERR(mr)) {
			CERROR("Failed ib_reg_phys_mr addr %#llx size %#llx : %ld\n",
			       ipb.addr, ipb.size, PTR_ERR(mr));
			kiblnd_hdev_cleanup_mrs(hdev);
			return PTR_ERR(mr);
		}

		LASSERT(iova == ipb.addr);

		hdev->ibh_mrs[i] = mr;
	}

out:
	if (hdev->ibh_mr_size != ~0ULL || hdev->ibh_nmrs != 1)
		LCONSOLE_INFO("Register global MR array, MR size: %#llx, array size: %d\n",
			      hdev->ibh_mr_size, hdev->ibh_nmrs);
	return 0;
	return 0;
}
}


@@ -2564,14 +2342,9 @@ int kiblnd_dev_failover(kib_dev_t *dev)
			kiblnd_fail_poolset(&net->ibn_tx_ps[i]->tps_poolset,
			kiblnd_fail_poolset(&net->ibn_tx_ps[i]->tps_poolset,
					    &zombie_tpo);
					    &zombie_tpo);


			if (net->ibn_fmr_ps != NULL) {
			if (net->ibn_fmr_ps)
				kiblnd_fail_fmr_poolset(net->ibn_fmr_ps[i],
				kiblnd_fail_fmr_poolset(net->ibn_fmr_ps[i],
							&zombie_fpo);
							&zombie_fpo);

			} else if (net->ibn_pmr_ps != NULL) {
				kiblnd_fail_poolset(&net->ibn_pmr_ps[i]->
						    pps_poolset, &zombie_ppo);
			}
		}
		}
	}
	}


+1 −31
Original line number Original line Diff line number Diff line
@@ -104,7 +104,6 @@ typedef struct {
	int          *kib_map_on_demand;     /* map-on-demand if RD has more
	int          *kib_map_on_demand;     /* map-on-demand if RD has more
					      * fragments than this value, 0
					      * fragments than this value, 0
					      * disable map-on-demand */
					      * disable map-on-demand */
	int          *kib_pmr_pool_size;     /* # physical MR in pool */
	int          *kib_fmr_pool_size;     /* # FMRs in pool */
	int          *kib_fmr_pool_size;     /* # FMRs in pool */
	int          *kib_fmr_flush_trigger; /* When to trigger FMR flush */
	int          *kib_fmr_flush_trigger; /* When to trigger FMR flush */
	int          *kib_fmr_cache;         /* enable FMR pool cache? */
	int          *kib_fmr_cache;         /* enable FMR pool cache? */
@@ -163,7 +162,6 @@ kiblnd_concurrent_sends_v1(void)
/* Pools (shared by connections on each CPT) */
/* Pools (shared by connections on each CPT) */
/* These pools can grow at runtime, so don't need give a very large value */
/* These pools can grow at runtime, so don't need give a very large value */
#define IBLND_TX_POOL			256
#define IBLND_TX_POOL			256
#define IBLND_PMR_POOL			256
#define IBLND_FMR_POOL			256
#define IBLND_FMR_POOL			256
#define IBLND_FMR_POOL_FLUSH		192
#define IBLND_FMR_POOL_FLUSH		192


@@ -232,17 +230,6 @@ typedef struct {
	struct page        *ibp_pages[0];       /* page array */
	struct page        *ibp_pages[0];       /* page array */
} kib_pages_t;
} kib_pages_t;


struct kib_pmr_pool;

typedef struct {
	struct list_head    pmr_list;           /* chain node */
	struct ib_phys_buf  *pmr_ipb;           /* physical buffer */
	struct ib_mr        *pmr_mr;            /* IB MR */
	struct kib_pmr_pool *pmr_pool;          /* owner of this MR */
	__u64               pmr_iova;           /* Virtual I/O address */
	int                 pmr_refcount;       /* reference count */
} kib_phys_mr_t;

struct kib_pool;
struct kib_pool;
struct kib_poolset;
struct kib_poolset;


@@ -298,15 +285,6 @@ typedef struct {
	kib_pages_t           *tpo_tx_pages;      /* premapped tx msg pages */
	kib_pages_t           *tpo_tx_pages;      /* premapped tx msg pages */
} kib_tx_pool_t;
} kib_tx_pool_t;


typedef struct {
	kib_poolset_t         pps_poolset;        /* pool-set */
} kib_pmr_poolset_t;

typedef struct kib_pmr_pool {
	struct kib_hca_dev    *ppo_hdev;          /* device for this pool */
	kib_pool_t            ppo_pool;           /* pool */
} kib_pmr_pool_t;

typedef struct {
typedef struct {
	spinlock_t            fps_lock;            /* serialize */
	spinlock_t            fps_lock;            /* serialize */
	struct kib_net        *fps_net;            /* IB network */
	struct kib_net        *fps_net;            /* IB network */
@@ -347,7 +325,6 @@ typedef struct kib_net {


	kib_tx_poolset_t      **ibn_tx_ps;    /* tx pool-set */
	kib_tx_poolset_t      **ibn_tx_ps;    /* tx pool-set */
	kib_fmr_poolset_t     **ibn_fmr_ps;   /* fmr pool-set */
	kib_fmr_poolset_t     **ibn_fmr_ps;   /* fmr pool-set */
	kib_pmr_poolset_t     **ibn_pmr_ps;   /* pmr pool-set */


	kib_dev_t             *ibn_dev;       /* underlying IB device */
	kib_dev_t             *ibn_dev;       /* underlying IB device */
} kib_net_t;
} kib_net_t;
@@ -554,10 +531,7 @@ typedef struct kib_tx /* transmit message */
	int                    tx_nfrags;     /* # entries in... */
	int                    tx_nfrags;     /* # entries in... */
	struct scatterlist     *tx_frags;     /* dma_map_sg descriptor */
	struct scatterlist     *tx_frags;     /* dma_map_sg descriptor */
	__u64                  *tx_pages;     /* rdma phys page addrs */
	__u64                  *tx_pages;     /* rdma phys page addrs */
	union {
		kib_phys_mr_t  *pmr;          /* MR for physical buffer */
	kib_fmr_t		fmr;	      /* FMR */
	kib_fmr_t		fmr;	      /* FMR */
	}                      tx_u;
	int                    tx_dmadir;     /* dma direction */
	int                    tx_dmadir;     /* dma direction */
} kib_tx_t;
} kib_tx_t;


@@ -978,10 +952,6 @@ int kiblnd_fmr_pool_map(kib_fmr_poolset_t *fps, __u64 *pages,
			 int npages, __u64 iov, kib_fmr_t *fmr);
			 int npages, __u64 iov, kib_fmr_t *fmr);
void kiblnd_fmr_pool_unmap(kib_fmr_t *fmr, int status);
void kiblnd_fmr_pool_unmap(kib_fmr_t *fmr, int status);


int  kiblnd_pmr_pool_map(kib_pmr_poolset_t *pps, kib_hca_dev_t *hdev,
			 kib_rdma_desc_t *rd, __u64 *iova, kib_phys_mr_t **pp_pmr);
void kiblnd_pmr_pool_unmap(kib_phys_mr_t *pmr);

int  kiblnd_startup (lnet_ni_t *ni);
int  kiblnd_startup (lnet_ni_t *ni);
void kiblnd_shutdown (lnet_ni_t *ni);
void kiblnd_shutdown (lnet_ni_t *ni);
int  kiblnd_ctl (lnet_ni_t *ni, unsigned int cmd, void *arg);
int  kiblnd_ctl (lnet_ni_t *ni, unsigned int cmd, void *arg);
+6 −49
Original line number Original line Diff line number Diff line
@@ -121,7 +121,6 @@ kiblnd_get_idle_tx(lnet_ni_t *ni, lnet_nid_t target)
	LASSERT(tx->tx_conn == NULL);
	LASSERT(tx->tx_conn == NULL);
	LASSERT(tx->tx_lntmsg[0] == NULL);
	LASSERT(tx->tx_lntmsg[0] == NULL);
	LASSERT(tx->tx_lntmsg[1] == NULL);
	LASSERT(tx->tx_lntmsg[1] == NULL);
	LASSERT(tx->tx_u.pmr == NULL);
	LASSERT(tx->tx_nfrags == 0);
	LASSERT(tx->tx_nfrags == 0);


	return tx;
	return tx;
@@ -575,7 +574,7 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
	cpt = tx->tx_pool->tpo_pool.po_owner->ps_cpt;
	cpt = tx->tx_pool->tpo_pool.po_owner->ps_cpt;


	fps = net->ibn_fmr_ps[cpt];
	fps = net->ibn_fmr_ps[cpt];
	rc = kiblnd_fmr_pool_map(fps, pages, npages, 0, &tx->tx_u.fmr);
	rc = kiblnd_fmr_pool_map(fps, pages, npages, 0, &tx->fmr);
	if (rc != 0) {
	if (rc != 0) {
		CERROR("Can't map %d pages: %d\n", npages, rc);
		CERROR("Can't map %d pages: %d\n", npages, rc);
		return rc;
		return rc;
@@ -583,8 +582,8 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)


	/* If rd is not tx_rd, it's going to get sent to a peer, who will need
	/* If rd is not tx_rd, it's going to get sent to a peer, who will need
	 * the rkey */
	 * the rkey */
	rd->rd_key = (rd != tx->tx_rd) ? tx->tx_u.fmr.fmr_pfmr->fmr->rkey :
	rd->rd_key = (rd != tx->tx_rd) ? tx->fmr.fmr_pfmr->fmr->rkey :
					 tx->tx_u.fmr.fmr_pfmr->fmr->lkey;
					 tx->fmr.fmr_pfmr->fmr->lkey;
	rd->rd_frags[0].rf_addr &= ~hdev->ibh_page_mask;
	rd->rd_frags[0].rf_addr &= ~hdev->ibh_page_mask;
	rd->rd_frags[0].rf_nob = nob;
	rd->rd_frags[0].rf_nob = nob;
	rd->rd_nfrags = 1;
	rd->rd_nfrags = 1;
@@ -592,42 +591,6 @@ kiblnd_fmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
	return 0;
	return 0;
}
}


static int
kiblnd_pmr_map_tx(kib_net_t *net, kib_tx_t *tx, kib_rdma_desc_t *rd, int nob)
{
	kib_hca_dev_t *hdev;
	kib_pmr_poolset_t *pps;
	__u64 iova;
	int cpt;
	int rc;

	LASSERT(tx->tx_pool != NULL);
	LASSERT(tx->tx_pool->tpo_pool.po_owner != NULL);

	hdev = tx->tx_pool->tpo_hdev;

	iova = rd->rd_frags[0].rf_addr & ~hdev->ibh_page_mask;

	cpt = tx->tx_pool->tpo_pool.po_owner->ps_cpt;

	pps = net->ibn_pmr_ps[cpt];
	rc = kiblnd_pmr_pool_map(pps, hdev, rd, &iova, &tx->tx_u.pmr);
	if (rc != 0) {
		CERROR("Failed to create MR by phybuf: %d\n", rc);
		return rc;
	}

	/* If rd is not tx_rd, it's going to get sent to a peer, who will need
	 * the rkey */
	rd->rd_key = (rd != tx->tx_rd) ? tx->tx_u.pmr->pmr_mr->rkey :
					 tx->tx_u.pmr->pmr_mr->lkey;
	rd->rd_nfrags = 1;
	rd->rd_frags[0].rf_addr = iova;
	rd->rd_frags[0].rf_nob = nob;

	return 0;
}

void
void
kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx)
kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx)
{
{
@@ -635,13 +598,9 @@ kiblnd_unmap_tx(lnet_ni_t *ni, kib_tx_t *tx)


	LASSERT(net != NULL);
	LASSERT(net != NULL);


	if (net->ibn_fmr_ps != NULL && tx->tx_u.fmr.fmr_pfmr != NULL) {
	if (net->ibn_fmr_ps && tx->fmr.fmr_pfmr) {
		kiblnd_fmr_pool_unmap(&tx->tx_u.fmr, tx->tx_status);
		kiblnd_fmr_pool_unmap(&tx->fmr, tx->tx_status);
		tx->tx_u.fmr.fmr_pfmr = NULL;
		tx->fmr.fmr_pfmr = NULL;

	} else if (net->ibn_pmr_ps != NULL && tx->tx_u.pmr != NULL) {
		kiblnd_pmr_pool_unmap(tx->tx_u.pmr);
		tx->tx_u.pmr = NULL;
	}
	}


	if (tx->tx_nfrags != 0) {
	if (tx->tx_nfrags != 0) {
@@ -687,8 +646,6 @@ kiblnd_map_tx(lnet_ni_t *ni, kib_tx_t *tx,


	if (net->ibn_fmr_ps != NULL)
	if (net->ibn_fmr_ps != NULL)
		return kiblnd_fmr_map_tx(net, tx, rd, nob);
		return kiblnd_fmr_map_tx(net, tx, rd, nob);
	else if (net->ibn_pmr_ps != NULL)
		return kiblnd_pmr_map_tx(net, tx, rd, nob);


	return -EINVAL;
	return -EINVAL;
}
}
+0 −6
Original line number Original line Diff line number Diff line
@@ -126,11 +126,6 @@ static int fmr_cache = 1;
module_param(fmr_cache, int, 0444);
module_param(fmr_cache, int, 0444);
MODULE_PARM_DESC(fmr_cache, "non-zero to enable FMR caching");
MODULE_PARM_DESC(fmr_cache, "non-zero to enable FMR caching");


/* NB: this value is shared by all CPTs, it can grow at runtime */
static int pmr_pool_size = 512;
module_param(pmr_pool_size, int, 0444);
MODULE_PARM_DESC(pmr_pool_size, "size of MR cache pmr pool on each CPT");

/*
/*
 * 0: disable failover
 * 0: disable failover
 * 1: enable failover if necessary
 * 1: enable failover if necessary
@@ -170,7 +165,6 @@ kib_tunables_t kiblnd_tunables = {
	.kib_fmr_pool_size     = &fmr_pool_size,
	.kib_fmr_pool_size     = &fmr_pool_size,
	.kib_fmr_flush_trigger = &fmr_flush_trigger,
	.kib_fmr_flush_trigger = &fmr_flush_trigger,
	.kib_fmr_cache         = &fmr_cache,
	.kib_fmr_cache         = &fmr_cache,
	.kib_pmr_pool_size     = &pmr_pool_size,
	.kib_require_priv_port = &require_privileged_port,
	.kib_require_priv_port = &require_privileged_port,
	.kib_use_priv_port     = &use_privileged_port,
	.kib_use_priv_port     = &use_privileged_port,
	.kib_nscheds           = &nscheds
	.kib_nscheds           = &nscheds