Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit c534a107 authored by Tejun Heo's avatar Tejun Heo
Browse files

rds/ib: use system_wq instead of rds_ib_fmr_wq



With cmwq, there's no reason to use dedicated rds_ib_fmr_wq - it's not
in the memory reclaim path and the maximum number of concurrent work
items is bound by the number of devices.  Drop it and use system_wq
instead.  This rds_ib_fmr_init/exit() noops.  Both removed.

Signed-off-by: default avatarTejun Heo <tj@kernel.org>
Cc: Andy Grover <andy.grover@oracle.com>
parent aa70c585
Loading
Loading
Loading
Loading
+1 −8
Original line number Diff line number Diff line
@@ -364,7 +364,6 @@ void rds_ib_exit(void)
	rds_ib_sysctl_exit();
	rds_ib_recv_exit();
	rds_trans_unregister(&rds_ib_transport);
	rds_ib_fmr_exit();
}

struct rds_transport rds_ib_transport = {
@@ -400,13 +399,9 @@ int rds_ib_init(void)

	INIT_LIST_HEAD(&rds_ib_devices);

	ret = rds_ib_fmr_init();
	if (ret)
		goto out;

	ret = ib_register_client(&rds_ib_client);
	if (ret)
		goto out_fmr_exit;
		goto out;

	ret = rds_ib_sysctl_init();
	if (ret)
@@ -430,8 +425,6 @@ int rds_ib_init(void)
	rds_ib_sysctl_exit();
out_ibreg:
	rds_ib_unregister_client();
out_fmr_exit:
	rds_ib_fmr_exit();
out:
	return ret;
}
+0 −2
Original line number Diff line number Diff line
@@ -307,8 +307,6 @@ void *rds_ib_get_mr(struct scatterlist *sg, unsigned long nents,
void rds_ib_sync_mr(void *trans_private, int dir);
void rds_ib_free_mr(void *trans_private, int invalidate);
void rds_ib_flush_mrs(void);
int rds_ib_fmr_init(void);
void rds_ib_fmr_exit(void);

/* ib_recv.c */
int rds_ib_recv_init(void);
+3 −24
Original line number Diff line number Diff line
@@ -38,8 +38,6 @@
#include "ib.h"
#include "xlist.h"

static struct workqueue_struct *rds_ib_fmr_wq;

static DEFINE_PER_CPU(unsigned long, clean_list_grace);
#define CLEAN_LIST_BUSY_BIT 0

@@ -307,7 +305,7 @@ static struct rds_ib_mr *rds_ib_alloc_fmr(struct rds_ib_device *rds_ibdev)
	int err = 0, iter = 0;

	if (atomic_read(&pool->dirty_count) >= pool->max_items / 10)
		queue_delayed_work(rds_ib_fmr_wq, &pool->flush_worker, 10);
		schedule_delayed_work(&pool->flush_worker, 10);

	while (1) {
		ibmr = rds_ib_reuse_fmr(pool);
@@ -696,24 +694,6 @@ static int rds_ib_flush_mr_pool(struct rds_ib_mr_pool *pool,
	return ret;
}

int rds_ib_fmr_init(void)
{
	rds_ib_fmr_wq = create_workqueue("rds_fmr_flushd");
	if (!rds_ib_fmr_wq)
		return -ENOMEM;
	return 0;
}

/*
 * By the time this is called all the IB devices should have been torn down and
 * had their pools freed.  As each pool is freed its work struct is waited on,
 * so the pool flushing work queue should be idle by the time we get here.
 */
void rds_ib_fmr_exit(void)
{
	destroy_workqueue(rds_ib_fmr_wq);
}

static void rds_ib_mr_pool_flush_worker(struct work_struct *work)
{
	struct rds_ib_mr_pool *pool = container_of(work, struct rds_ib_mr_pool, flush_worker.work);
@@ -741,7 +721,7 @@ void rds_ib_free_mr(void *trans_private, int invalidate)
	/* If we've pinned too many pages, request a flush */
	if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
	    atomic_read(&pool->dirty_count) >= pool->max_items / 10)
		queue_delayed_work(rds_ib_fmr_wq, &pool->flush_worker, 10);
		schedule_delayed_work(&pool->flush_worker, 10);

	if (invalidate) {
		if (likely(!in_interrupt())) {
@@ -749,8 +729,7 @@ void rds_ib_free_mr(void *trans_private, int invalidate)
		} else {
			/* We get here if the user created a MR marked
			 * as use_once and invalidate at the same time. */
			queue_delayed_work(rds_ib_fmr_wq,
					   &pool->flush_worker, 10);
			schedule_delayed_work(&pool->flush_worker, 10);
		}
	}