Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 856cc4c2 authored by Mike Marciniszyn's avatar Mike Marciniszyn Committed by Doug Ledford
Browse files

IB/hfi1: Add the capability for reserved operations



This fix allows for support of in-kernel reserved operations
without impacting the ULP user.

The low level driver can register a non-zero value which
will be transparently added to the send queue size and hidden
from the ULP in every respect.

ULP post sends will never see a full queue due to a reserved
post send and reserved operations will never exceed that
registered value.

The s_avail will continue to track the ULP swqe availability
and the difference between the reserved value and the reserved
in use will track reserved availabity.

Reviewed-by: default avatarAshutosh Dixit <ashutosh.dixit@intel.com>
Signed-off-by: default avatarMike Marciniszyn <mike.marciniszyn@intel.com>
Signed-off-by: default avatarDennis Dalessandro <dennis.dalessandro@intel.com>
Signed-off-by: default avatarDoug Ledford <dledford@redhat.com>
parent 23002d5b
Loading
Loading
Loading
Loading
+62 −23
Original line number Original line Diff line number Diff line
@@ -584,6 +584,7 @@ static void rvt_reset_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp,
		qp->r_rq.wq->tail = 0;
		qp->r_rq.wq->tail = 0;
	}
	}
	qp->r_sge.num_sge = 0;
	qp->r_sge.num_sge = 0;
	atomic_set(&qp->s_reserved_used, 0);
}
}


/**
/**
@@ -645,7 +646,8 @@ struct ib_qp *rvt_create_qp(struct ib_pd *ibpd,
			return ERR_PTR(-EINVAL);
			return ERR_PTR(-EINVAL);
	}
	}
	sqsize =
	sqsize =
		init_attr->cap.max_send_wr + 1;
		init_attr->cap.max_send_wr + 1 +
		rdi->dparms.reserved_operations;
	switch (init_attr->qp_type) {
	switch (init_attr->qp_type) {
	case IB_QPT_SMI:
	case IB_QPT_SMI:
	case IB_QPT_GSI:
	case IB_QPT_GSI:
@@ -1335,7 +1337,8 @@ int rvt_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
	attr->sq_psn = qp->s_next_psn & rdi->dparms.psn_mask;
	attr->sq_psn = qp->s_next_psn & rdi->dparms.psn_mask;
	attr->dest_qp_num = qp->remote_qpn;
	attr->dest_qp_num = qp->remote_qpn;
	attr->qp_access_flags = qp->qp_access_flags;
	attr->qp_access_flags = qp->qp_access_flags;
	attr->cap.max_send_wr = qp->s_size - 1;
	attr->cap.max_send_wr = qp->s_size - 1 -
		rdi->dparms.reserved_operations;
	attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1;
	attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1;
	attr->cap.max_send_sge = qp->s_max_sge;
	attr->cap.max_send_sge = qp->s_max_sge;
	attr->cap.max_recv_sge = qp->r_rq.max_sge;
	attr->cap.max_recv_sge = qp->r_rq.max_sge;
@@ -1494,27 +1497,65 @@ static inline int rvt_qp_valid_operation(
}
}


/**
/**
 * qp_get_savail - return number of avail send entries
 * rvt_qp_is_avail - determine queue capacity
 * @qp - the qp
 * @qp - the qp
 * @rdi - the rdmavt device
 * @reserved_op - is reserved operation
 *
 *
 * This assumes the s_hlock is held but the s_last
 * This assumes the s_hlock is held but the s_last
 * qp variable is uncontrolled.
 * qp variable is uncontrolled.
 *
 *
 * The return is adjusted to not count device specific
 * For non reserved operations, the qp->s_avail
 * reserved operations.
 * may be changed.
 *
 * The return value is zero or a -ENOMEM.
 */
 */
static inline u32 qp_get_savail(struct rvt_qp *qp)
static inline int rvt_qp_is_avail(
	struct rvt_qp *qp,
	struct rvt_dev_info *rdi,
	bool reserved_op)
{
{
	u32 slast;
	u32 slast;
	u32 ret;
	u32 avail;

	u32 reserved_used;

	/* see rvt_qp_wqe_unreserve() */
	smp_mb__before_atomic();
	reserved_used = atomic_read(&qp->s_reserved_used);
	if (unlikely(reserved_op)) {
		/* see rvt_qp_wqe_unreserve() */
		smp_mb__before_atomic();
		if (reserved_used >= rdi->dparms.reserved_operations)
			return -ENOMEM;
		return 0;
	}
	/* non-reserved operations */
	if (likely(qp->s_avail))
		return 0;
	smp_read_barrier_depends(); /* see rc.c */
	smp_read_barrier_depends(); /* see rc.c */
	slast = ACCESS_ONCE(qp->s_last);
	slast = ACCESS_ONCE(qp->s_last);
	if (qp->s_head >= slast)
	if (qp->s_head >= slast)
		ret = qp->s_size - (qp->s_head - slast);
		avail = qp->s_size - (qp->s_head - slast);
	else
	else
		ret = slast - qp->s_head;
		avail = slast - qp->s_head;
	return ret - 1;

	/* see rvt_qp_wqe_unreserve() */
	smp_mb__before_atomic();
	reserved_used = atomic_read(&qp->s_reserved_used);
	avail =  avail - 1 -
		(rdi->dparms.reserved_operations - reserved_used);
	/* insure we don't assign a negative s_avail */
	if ((s32)avail <= 0)
		return -ENOMEM;
	qp->s_avail = avail;
	if (WARN_ON(qp->s_avail >
		    (qp->s_size - 1 - rdi->dparms.reserved_operations)))
		rvt_pr_err(rdi,
			   "More avail entries than QP RB size.\nQP: %u, size: %u, avail: %u\nhead: %u, tail: %u, cur: %u, acked: %u, last: %u",
			   qp->ibqp.qp_num, qp->s_size, qp->s_avail,
			   qp->s_head, qp->s_tail, qp->s_cur,
			   qp->s_acked, qp->s_last);
	return 0;
}
}


/**
/**
@@ -1537,6 +1578,7 @@ static int rvt_post_one_wr(struct rvt_qp *qp,
	u8 log_pmtu;
	u8 log_pmtu;
	int ret;
	int ret;
	size_t cplen;
	size_t cplen;
	bool reserved_op;


	BUILD_BUG_ON(IB_QPT_MAX >= (sizeof(u32) * BITS_PER_BYTE));
	BUILD_BUG_ON(IB_QPT_MAX >= (sizeof(u32) * BITS_PER_BYTE));


@@ -1574,18 +1616,12 @@ static int rvt_post_one_wr(struct rvt_qp *qp,
		}
		}
	}
	}


	reserved_op = rdi->post_parms[wr->opcode].flags &
			RVT_OPERATION_USE_RESERVE;
	/* check for avail */
	/* check for avail */
	if (unlikely(!qp->s_avail)) {
	ret = rvt_qp_is_avail(qp, rdi, reserved_op);
		qp->s_avail = qp_get_savail(qp);
	if (ret)
		if (WARN_ON(qp->s_avail > (qp->s_size - 1)))
		return ret;
			rvt_pr_err(rdi,
				   "More avail entries than QP RB size.\nQP: %u, size: %u, avail: %u\nhead: %u, tail: %u, cur: %u, acked: %u, last: %u",
				   qp->ibqp.qp_num, qp->s_size, qp->s_avail,
				   qp->s_head, qp->s_tail, qp->s_cur,
				   qp->s_acked, qp->s_last);
		if (!qp->s_avail)
			return -ENOMEM;
	}
	next = qp->s_head + 1;
	next = qp->s_head + 1;
	if (next >= qp->s_size)
	if (next >= qp->s_size)
		next = 0;
		next = 0;
@@ -1653,8 +1689,11 @@ static int rvt_post_one_wr(struct rvt_qp *qp,
		qp->s_next_psn = wqe->lpsn + 1;
		qp->s_next_psn = wqe->lpsn + 1;
	}
	}
	trace_rvt_post_one_wr(qp, wqe);
	trace_rvt_post_one_wr(qp, wqe);
	smp_wmb(); /* see request builders */
	if (unlikely(reserved_op))
		rvt_qp_wqe_reserve(qp, wqe);
	else
		qp->s_avail--;
		qp->s_avail--;
	smp_wmb(); /* see request builders */
	qp->s_head = next;
	qp->s_head = next;


	return 0;
	return 0;
+1 −0
Original line number Original line Diff line number Diff line
@@ -158,6 +158,7 @@ struct rvt_driver_params {
	u32 max_mad_size;
	u32 max_mad_size;
	u8 qos_shift;
	u8 qos_shift;
	u8 max_rdma_atomic;
	u8 max_rdma_atomic;
	u8 reserved_operations;
};
};


/* Protection domain */
/* Protection domain */
+50 −0
Original line number Original line Diff line number Diff line
@@ -144,6 +144,11 @@
#define RVT_PROCESS_OR_FLUSH_SEND \
#define RVT_PROCESS_OR_FLUSH_SEND \
	(RVT_PROCESS_SEND_OK | RVT_FLUSH_SEND)
	(RVT_PROCESS_SEND_OK | RVT_FLUSH_SEND)


/*
 * Internal send flags
 */
#define RVT_SEND_RESERVE_USED           IB_SEND_RESERVED_START

/*
/*
 * Send work request queue entry.
 * Send work request queue entry.
 * The size of the sg_list is determined when the QP is created and stored
 * The size of the sg_list is determined when the QP is created and stored
@@ -232,6 +237,7 @@ struct rvt_ack_entry {
#define RVT_OPERATION_ATOMIC      0x00000002
#define RVT_OPERATION_ATOMIC      0x00000002
#define RVT_OPERATION_ATOMIC_SGE  0x00000004
#define RVT_OPERATION_ATOMIC_SGE  0x00000004
#define RVT_OPERATION_LOCAL       0x00000008
#define RVT_OPERATION_LOCAL       0x00000008
#define RVT_OPERATION_USE_RESERVE 0x00000010


#define RVT_OPERATION_MAX (IB_WR_RESERVED10 + 1)
#define RVT_OPERATION_MAX (IB_WR_RESERVED10 + 1)


@@ -328,6 +334,7 @@ struct rvt_qp {
	u32 s_next_psn;         /* PSN for next request */
	u32 s_next_psn;         /* PSN for next request */
	u32 s_avail;            /* number of entries avail */
	u32 s_avail;            /* number of entries avail */
	u32 s_ssn;              /* SSN of tail entry */
	u32 s_ssn;              /* SSN of tail entry */
	atomic_t s_reserved_used; /* reserved entries in use */


	spinlock_t s_lock ____cacheline_aligned_in_smp;
	spinlock_t s_lock ____cacheline_aligned_in_smp;
	u32 s_flags;
	u32 s_flags;
@@ -459,6 +466,49 @@ static inline struct rvt_rwqe *rvt_get_rwqe_ptr(struct rvt_rq *rq, unsigned n)
		  rq->max_sge * sizeof(struct ib_sge)) * n);
		  rq->max_sge * sizeof(struct ib_sge)) * n);
}
}


/**
 * rvt_qp_wqe_reserve - reserve operation
 * @qp - the rvt qp
 * @wqe - the send wqe
 *
 * This routine used in post send to record
 * a wqe relative reserved operation use.
 */
static inline void rvt_qp_wqe_reserve(
	struct rvt_qp *qp,
	struct rvt_swqe *wqe)
{
	wqe->wr.send_flags |= RVT_SEND_RESERVE_USED;
	atomic_inc(&qp->s_reserved_used);
}

/**
 * rvt_qp_wqe_unreserve - clean reserved operation
 * @qp - the rvt qp
 * @wqe - the send wqe
 *
 * This decrements the reserve use count.
 *
 * This call MUST precede the change to
 * s_last to insure that post send sees a stable
 * s_avail.
 *
 * An smp_mp__after_atomic() is used to insure
 * the compiler does not juggle the order of the s_last
 * ring index and the decrementing of s_reserved_used.
 */
static inline void rvt_qp_wqe_unreserve(
	struct rvt_qp *qp,
	struct rvt_swqe *wqe)
{
	if (unlikely(wqe->wr.send_flags & RVT_SEND_RESERVE_USED)) {
		wqe->wr.send_flags &= ~RVT_SEND_RESERVE_USED;
		atomic_dec(&qp->s_reserved_used);
		/* insure no compiler re-order up to s_last change */
		smp_mb__after_atomic();
	}
}

extern const int  ib_rvt_state_ops[];
extern const int  ib_rvt_state_ops[];


struct rvt_dev_info;
struct rvt_dev_info;