RDMA/hns: Use rdma_user_mmap_io
Rely on the new core code helper to map BAR memory from the driver. Signed-off-by: Jason Gunthorpe <jgg@mellanox.com> Signed-off-by: Leon Romanovsky <leonro@mellanox.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
This commit is contained in:
parent
e2cd1d1ad2
commit
6745d356ab
@ -219,19 +219,11 @@ struct hns_roce_uar {
|
|||||||
unsigned long logic_idx;
|
unsigned long logic_idx;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct hns_roce_vma_data {
|
|
||||||
struct list_head list;
|
|
||||||
struct vm_area_struct *vma;
|
|
||||||
struct mutex *vma_list_mutex;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct hns_roce_ucontext {
|
struct hns_roce_ucontext {
|
||||||
struct ib_ucontext ibucontext;
|
struct ib_ucontext ibucontext;
|
||||||
struct hns_roce_uar uar;
|
struct hns_roce_uar uar;
|
||||||
struct list_head page_list;
|
struct list_head page_list;
|
||||||
struct mutex page_mutex;
|
struct mutex page_mutex;
|
||||||
struct list_head vma_list;
|
|
||||||
struct mutex vma_list_mutex;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
struct hns_roce_pd {
|
struct hns_roce_pd {
|
||||||
|
@ -344,8 +344,6 @@ static struct ib_ucontext *hns_roce_alloc_ucontext(struct ib_device *ib_dev,
|
|||||||
if (ret)
|
if (ret)
|
||||||
goto error_fail_uar_alloc;
|
goto error_fail_uar_alloc;
|
||||||
|
|
||||||
INIT_LIST_HEAD(&context->vma_list);
|
|
||||||
mutex_init(&context->vma_list_mutex);
|
|
||||||
if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) {
|
if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) {
|
||||||
INIT_LIST_HEAD(&context->page_list);
|
INIT_LIST_HEAD(&context->page_list);
|
||||||
mutex_init(&context->page_mutex);
|
mutex_init(&context->page_mutex);
|
||||||
@ -376,76 +374,34 @@ static int hns_roce_dealloc_ucontext(struct ib_ucontext *ibcontext)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void hns_roce_vma_open(struct vm_area_struct *vma)
|
|
||||||
{
|
|
||||||
vma->vm_ops = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void hns_roce_vma_close(struct vm_area_struct *vma)
|
|
||||||
{
|
|
||||||
struct hns_roce_vma_data *vma_data;
|
|
||||||
|
|
||||||
vma_data = (struct hns_roce_vma_data *)vma->vm_private_data;
|
|
||||||
vma_data->vma = NULL;
|
|
||||||
mutex_lock(vma_data->vma_list_mutex);
|
|
||||||
list_del(&vma_data->list);
|
|
||||||
mutex_unlock(vma_data->vma_list_mutex);
|
|
||||||
kfree(vma_data);
|
|
||||||
}
|
|
||||||
|
|
||||||
static const struct vm_operations_struct hns_roce_vm_ops = {
|
|
||||||
.open = hns_roce_vma_open,
|
|
||||||
.close = hns_roce_vma_close,
|
|
||||||
};
|
|
||||||
|
|
||||||
static int hns_roce_set_vma_data(struct vm_area_struct *vma,
|
|
||||||
struct hns_roce_ucontext *context)
|
|
||||||
{
|
|
||||||
struct list_head *vma_head = &context->vma_list;
|
|
||||||
struct hns_roce_vma_data *vma_data;
|
|
||||||
|
|
||||||
vma_data = kzalloc(sizeof(*vma_data), GFP_KERNEL);
|
|
||||||
if (!vma_data)
|
|
||||||
return -ENOMEM;
|
|
||||||
|
|
||||||
vma_data->vma = vma;
|
|
||||||
vma_data->vma_list_mutex = &context->vma_list_mutex;
|
|
||||||
vma->vm_private_data = vma_data;
|
|
||||||
vma->vm_ops = &hns_roce_vm_ops;
|
|
||||||
|
|
||||||
mutex_lock(&context->vma_list_mutex);
|
|
||||||
list_add(&vma_data->list, vma_head);
|
|
||||||
mutex_unlock(&context->vma_list_mutex);
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int hns_roce_mmap(struct ib_ucontext *context,
|
static int hns_roce_mmap(struct ib_ucontext *context,
|
||||||
struct vm_area_struct *vma)
|
struct vm_area_struct *vma)
|
||||||
{
|
{
|
||||||
struct hns_roce_dev *hr_dev = to_hr_dev(context->device);
|
struct hns_roce_dev *hr_dev = to_hr_dev(context->device);
|
||||||
|
|
||||||
if (((vma->vm_end - vma->vm_start) % PAGE_SIZE) != 0)
|
switch (vma->vm_pgoff) {
|
||||||
return -EINVAL;
|
case 0:
|
||||||
|
return rdma_user_mmap_io(context, vma,
|
||||||
|
to_hr_ucontext(context)->uar.pfn,
|
||||||
|
PAGE_SIZE,
|
||||||
|
pgprot_noncached(vma->vm_page_prot));
|
||||||
|
|
||||||
if (vma->vm_pgoff == 0) {
|
/* vm_pgoff: 1 -- TPTR */
|
||||||
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
case 1:
|
||||||
if (io_remap_pfn_range(vma, vma->vm_start,
|
if (!hr_dev->tptr_dma_addr || !hr_dev->tptr_size)
|
||||||
to_hr_ucontext(context)->uar.pfn,
|
return -EINVAL;
|
||||||
PAGE_SIZE, vma->vm_page_prot))
|
/*
|
||||||
return -EAGAIN;
|
* FIXME: using io_remap_pfn_range on the dma address returned
|
||||||
} else if (vma->vm_pgoff == 1 && hr_dev->tptr_dma_addr &&
|
* by dma_alloc_coherent is totally wrong.
|
||||||
hr_dev->tptr_size) {
|
*/
|
||||||
/* vm_pgoff: 1 -- TPTR */
|
return rdma_user_mmap_io(context, vma,
|
||||||
if (io_remap_pfn_range(vma, vma->vm_start,
|
hr_dev->tptr_dma_addr >> PAGE_SHIFT,
|
||||||
hr_dev->tptr_dma_addr >> PAGE_SHIFT,
|
hr_dev->tptr_size,
|
||||||
hr_dev->tptr_size,
|
vma->vm_page_prot);
|
||||||
vma->vm_page_prot))
|
|
||||||
return -EAGAIN;
|
|
||||||
} else
|
|
||||||
return -EINVAL;
|
|
||||||
|
|
||||||
return hns_roce_set_vma_data(vma, to_hr_ucontext(context));
|
default:
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,
|
static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,
|
||||||
@ -471,21 +427,6 @@ static int hns_roce_port_immutable(struct ib_device *ib_dev, u8 port_num,
|
|||||||
|
|
||||||
static void hns_roce_disassociate_ucontext(struct ib_ucontext *ibcontext)
|
static void hns_roce_disassociate_ucontext(struct ib_ucontext *ibcontext)
|
||||||
{
|
{
|
||||||
struct hns_roce_ucontext *context = to_hr_ucontext(ibcontext);
|
|
||||||
struct hns_roce_vma_data *vma_data, *n;
|
|
||||||
struct vm_area_struct *vma;
|
|
||||||
|
|
||||||
mutex_lock(&context->vma_list_mutex);
|
|
||||||
list_for_each_entry_safe(vma_data, n, &context->vma_list, list) {
|
|
||||||
vma = vma_data->vma;
|
|
||||||
zap_vma_ptes(vma, vma->vm_start, PAGE_SIZE);
|
|
||||||
|
|
||||||
vma->vm_flags &= ~(VM_SHARED | VM_MAYSHARE);
|
|
||||||
vma->vm_ops = NULL;
|
|
||||||
list_del(&vma_data->list);
|
|
||||||
kfree(vma_data);
|
|
||||||
}
|
|
||||||
mutex_unlock(&context->vma_list_mutex);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void hns_roce_unregister_device(struct hns_roce_dev *hr_dev)
|
static void hns_roce_unregister_device(struct hns_roce_dev *hr_dev)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user