Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit b8533ecc authored by Sagi Grimberg's avatar Sagi Grimberg Committed by Doug Ledford
Browse files

IB/qib: Remove old FRWR API



No ULP uses it anymore, go ahead and remove it.

Signed-off-by: default avatarSagi Grimberg <sagig@mellanox.com>
Acked-by: default avatarChristoph Hellwig <hch@lst.de>
Signed-off-by: default avatarDoug Ledford <dledford@redhat.com>
parent d3cfd002
Loading
Loading
Loading
Loading
+0 −56
Original line number Diff line number Diff line
@@ -335,62 +335,6 @@ int qib_rkey_ok(struct qib_qp *qp, struct qib_sge *sge,
	return 0;
}

/*
 * Initialize the memory region specified by the work reqeust.
 */
int qib_fast_reg_mr(struct qib_qp *qp, struct ib_send_wr *send_wr)
{
	struct ib_fast_reg_wr *wr = fast_reg_wr(send_wr);
	struct qib_lkey_table *rkt = &to_idev(qp->ibqp.device)->lk_table;
	struct qib_pd *pd = to_ipd(qp->ibqp.pd);
	struct qib_mregion *mr;
	u32 rkey = wr->rkey;
	unsigned i, n, m;
	int ret = -EINVAL;
	unsigned long flags;
	u64 *page_list;
	size_t ps;

	spin_lock_irqsave(&rkt->lock, flags);
	if (pd->user || rkey == 0)
		goto bail;

	mr = rcu_dereference_protected(
		rkt->table[(rkey >> (32 - ib_qib_lkey_table_size))],
		lockdep_is_held(&rkt->lock));
	if (unlikely(mr == NULL || qp->ibqp.pd != mr->pd))
		goto bail;

	if (wr->page_list_len > mr->max_segs)
		goto bail;

	ps = 1UL << wr->page_shift;
	if (wr->length > ps * wr->page_list_len)
		goto bail;

	mr->user_base = wr->iova_start;
	mr->iova = wr->iova_start;
	mr->lkey = rkey;
	mr->length = wr->length;
	mr->access_flags = wr->access_flags;
	page_list = wr->page_list->page_list;
	m = 0;
	n = 0;
	for (i = 0; i < wr->page_list_len; i++) {
		mr->map[m]->segs[n].vaddr = (void *) page_list[i];
		mr->map[m]->segs[n].length = ps;
		if (++n == QIB_SEGSZ) {
			m++;
			n = 0;
		}
	}

	ret = 0;
bail:
	spin_unlock_irqrestore(&rkt->lock, flags);
	return ret;
}

/*
 * Initialize the memory region specified by the work request.
 */
+1 −31
Original line number Diff line number Diff line
@@ -324,7 +324,7 @@ int qib_dereg_mr(struct ib_mr *ibmr)

/*
 * Allocate a memory region usable with the
 * IB_WR_FAST_REG_MR send work request.
 * IB_WR_REG_MR send work request.
 *
 * Return the memory region on success, otherwise return an errno.
 */
@@ -375,36 +375,6 @@ int qib_map_mr_sg(struct ib_mr *ibmr,
	return ib_sg_to_pages(ibmr, sg, sg_nents, qib_set_page);
}

struct ib_fast_reg_page_list *
qib_alloc_fast_reg_page_list(struct ib_device *ibdev, int page_list_len)
{
	unsigned size = page_list_len * sizeof(u64);
	struct ib_fast_reg_page_list *pl;

	if (size > PAGE_SIZE)
		return ERR_PTR(-EINVAL);

	pl = kzalloc(sizeof(*pl), GFP_KERNEL);
	if (!pl)
		return ERR_PTR(-ENOMEM);

	pl->page_list = kzalloc(size, GFP_KERNEL);
	if (!pl->page_list)
		goto err_free;

	return pl;

err_free:
	kfree(pl);
	return ERR_PTR(-ENOMEM);
}

void qib_free_fast_reg_page_list(struct ib_fast_reg_page_list *pl)
{
	kfree(pl->page_list);
	kfree(pl);
}

/**
 * qib_alloc_fmr - allocate a fast memory region
 * @pd: the protection domain for this memory region
+0 −8
Original line number Diff line number Diff line
@@ -365,9 +365,6 @@ static int qib_post_one_send(struct qib_qp *qp, struct ib_send_wr *wr,
	if (wr->opcode == IB_WR_REG_MR) {
		if (qib_reg_mr(qp, reg_wr(wr)))
			goto bail_inval;
	} else if (wr->opcode == IB_WR_FAST_REG_MR) {
		if (qib_fast_reg_mr(qp, wr))
			goto bail_inval;
	} else if (qp->ibqp.qp_type == IB_QPT_UC) {
		if ((unsigned) wr->opcode >= IB_WR_RDMA_READ)
			goto bail_inval;
@@ -407,9 +404,6 @@ static int qib_post_one_send(struct qib_qp *qp, struct ib_send_wr *wr,
	else if (wr->opcode == IB_WR_REG_MR)
		memcpy(&wqe->reg_wr, reg_wr(wr),
			sizeof(wqe->reg_wr));
	else if (wr->opcode == IB_WR_FAST_REG_MR)
		memcpy(&wqe->fast_reg_wr, fast_reg_wr(wr),
			sizeof(wqe->fast_reg_wr));
	else if (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM ||
		 wr->opcode == IB_WR_RDMA_WRITE ||
		 wr->opcode == IB_WR_RDMA_READ)
@@ -2267,8 +2261,6 @@ int qib_register_ib_device(struct qib_devdata *dd)
	ibdev->dereg_mr = qib_dereg_mr;
	ibdev->alloc_mr = qib_alloc_mr;
	ibdev->map_mr_sg = qib_map_mr_sg;
	ibdev->alloc_fast_reg_page_list = qib_alloc_fast_reg_page_list;
	ibdev->free_fast_reg_page_list = qib_free_fast_reg_page_list;
	ibdev->alloc_fmr = qib_alloc_fmr;
	ibdev->map_phys_fmr = qib_map_phys_fmr;
	ibdev->unmap_fmr = qib_unmap_fmr;
+0 −7
Original line number Diff line number Diff line
@@ -344,7 +344,6 @@ struct qib_swqe {
		struct ib_send_wr wr;   /* don't use wr.sg_list */
		struct ib_ud_wr ud_wr;
		struct ib_reg_wr reg_wr;
		struct ib_fast_reg_wr fast_reg_wr;
		struct ib_rdma_wr rdma_wr;
		struct ib_atomic_wr atomic_wr;
	};
@@ -1051,12 +1050,6 @@ int qib_map_mr_sg(struct ib_mr *ibmr,
		  struct scatterlist *sg,
		  int sg_nents);

struct ib_fast_reg_page_list *qib_alloc_fast_reg_page_list(
				struct ib_device *ibdev, int page_list_len);

void qib_free_fast_reg_page_list(struct ib_fast_reg_page_list *pl);

int qib_fast_reg_mr(struct qib_qp *qp, struct ib_send_wr *wr);
int qib_reg_mr(struct qib_qp *qp, struct ib_reg_wr *wr);

struct ib_fmr *qib_alloc_fmr(struct ib_pd *pd, int mr_access_flags,