Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 6745d356 authored by Jason Gunthorpe's avatar Jason Gunthorpe Committed by Doug Ledford
Browse files

RDMA/hns: Use rdma_user_mmap_io



Rely on the new core code helper to map BAR memory from the driver.

Signed-off-by: default avatarJason Gunthorpe <jgg@mellanox.com>
Signed-off-by: default avatarLeon Romanovsky <leonro@mellanox.com>
Signed-off-by: default avatarDoug Ledford <dledford@redhat.com>
parent e2cd1d1a
Loading
Loading
Loading
Loading
+0 −8
Original line number Original line Diff line number Diff line
@@ -219,19 +219,11 @@ struct hns_roce_uar {
	unsigned long	logic_idx;
	unsigned long	logic_idx;
};
};


struct hns_roce_vma_data {
	struct list_head list;
	struct vm_area_struct *vma;
	struct mutex *vma_list_mutex;
};

struct hns_roce_ucontext {
struct hns_roce_ucontext {
	struct ib_ucontext	ibucontext;
	struct ib_ucontext	ibucontext;
	struct hns_roce_uar	uar;
	struct hns_roce_uar	uar;
	struct list_head	page_list;
	struct list_head	page_list;
	struct mutex		page_mutex;
	struct mutex		page_mutex;
	struct list_head	vma_list;
	struct mutex		vma_list_mutex;
};
};


struct hns_roce_pd {
struct hns_roce_pd {
+21 −80
Original line number Original line Diff line number Diff line
@@ -344,8 +344,6 @@ static struct ib_ucontext *hns_roce_alloc_ucontext(struct ib_device *ib_dev,
	if (ret)
	if (ret)
		goto error_fail_uar_alloc;
		goto error_fail_uar_alloc;


	INIT_LIST_HEAD(&context->vma_list);
	mutex_init(&context->vma_list_mutex);
	if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) {
	if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) {
		INIT_LIST_HEAD(&context->page_list);
		INIT_LIST_HEAD(&context->page_list);
		mutex_init(&context->page_mutex);
		mutex_init(&context->page_mutex);
@@ -376,76 +374,34 @@ static int hns_roce_dealloc_ucontext(struct ib_ucontext *ibcontext)
	return 0;
	return 0;
}
}


static void hns_roce_vma_open(struct vm_area_struct *vma)
{
	vma->vm_ops = NULL;
}

static void hns_roce_vma_close(struct vm_area_struct *vma)
{
	struct hns_roce_vma_data *vma_data;

	vma_data = (struct hns_roce_vma_data *)vma->vm_private_data;
	vma_data->vma = NULL;
	mutex_lock(vma_data->vma_list_mutex);
	list_del(&vma_data->list);
	mutex_unlock(vma_data->vma_list_mutex);
	kfree(vma_data);
}

static const struct vm_operations_struct hns_roce_vm_ops = {
	.open = hns_roce_vma_open,
	.close = hns_roce_vma_close,
};

static int hns_roce_set_vma_data(struct vm_area_struct *vma,
				 struct hns_roce_ucontext *context)
{
	struct list_head *vma_head = &context->vma_list;
	struct hns_roce_vma_data *vma_data;

	vma_data = kzalloc(sizeof(*vma_data), GFP_KERNEL);
	if (!vma_data)
		return -ENOMEM;

	vma_data->vma = vma;
	vma_data->vma_list_mutex = &context->vma_list_mutex;
	vma->vm_private_data = vma_data;
	vma->vm_ops = &hns_roce_vm_ops;

	mutex_lock(&context->vma_list_mutex);
	list_add(&vma_data->list, vma_head);
	mutex_unlock(&context->vma_list_mutex);

	return 0;
}

static int hns_roce_mmap(struct ib_ucontext *context,
static int hns_roce_mmap(struct ib_ucontext *context,
			 struct vm_area_struct *vma)
			 struct vm_area_struct *vma)
{
{
	struct hns_roce_dev *hr_dev = to_hr_dev(context->device);
	struct hns_roce_dev *hr_dev = to_hr_dev(context->device);


	if (((vma->vm_end - vma->vm_start) % PAGE_SIZE) != 0)
	switch (vma->vm_pgoff) {
		return -EINVAL;
	case 0:

		return rdma_user_mmap_io(context, vma,
	if (vma->vm_pgoff == 0) {
		vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
		if (io_remap_pfn_range(vma, vma->vm_start,
					 to_hr_ucontext(context)->uar.pfn,
					 to_hr_ucontext(context)->uar.pfn,
				       PAGE_SIZE, vma->vm_page_prot))
					 PAGE_SIZE,
			return -EAGAIN;
					 pgprot_noncached(vma->vm_page_prot));
	} else if (vma->vm_pgoff == 1 && hr_dev->tptr_dma_addr &&

		   hr_dev->tptr_size) {
	/* vm_pgoff: 1 -- TPTR */
	/* vm_pgoff: 1 -- TPTR */
		if (io_remap_pfn_range(vma, vma->vm_start,
	case 1:
		if (!hr_dev->tptr_dma_addr || !hr_dev->tptr_size)
			return -EINVAL;
		/*
		 * FIXME: using io_remap_pfn_range on the dma address returned
		 * by dma_alloc_coherent is totally wrong.
		 */
		return rdma_user_mmap_io(context, vma,
					 hr_dev->tptr_dma_addr >> PAGE_SHIFT,
					 hr_dev->tptr_dma_addr >> PAGE_SHIFT,
					 hr_dev->tptr_size,
					 hr_dev->tptr_size,
				       vma->vm_page_prot))
					 vma->vm_page_prot);
			return -EAGAIN;
	} else
		return -EINVAL;


	return hns_roce_set_vma_data(vma, to_hr_ucontext(context));
	default:
		return -EINVAL;
	}
}
}


static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,
static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,
@@ -471,21 +427,6 @@ static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,


static void hns_roce_disassociate_ucontext(struct ib_ucontext *ibcontext)
static void hns_roce_disassociate_ucontext(struct ib_ucontext *ibcontext)
{
{
	struct hns_roce_ucontext *context = to_hr_ucontext(ibcontext);
	struct hns_roce_vma_data *vma_data, *n;
	struct vm_area_struct *vma;

	mutex_lock(&context->vma_list_mutex);
	list_for_each_entry_safe(vma_data, n, &context->vma_list, list) {
		vma = vma_data->vma;
		zap_vma_ptes(vma, vma->vm_start, PAGE_SIZE);

		vma->vm_flags &= ~(VM_SHARED | VM_MAYSHARE);
		vma->vm_ops = NULL;
		list_del(&vma_data->list);
		kfree(vma_data);
	}
	mutex_unlock(&context->vma_list_mutex);
}
}


static void hns_roce_unregister_device(struct hns_roce_dev *hr_dev)
static void hns_roce_unregister_device(struct hns_roce_dev *hr_dev)