Refactor the logic in io_register_pbuf_ring() into generic helpers: - io_copy_and_validate_buf_reg(): Copy out user arg and validate user arg and buffer registration parameters - io_alloc_new_buffer_list(): Allocate and initialize a new buffer list for the given buffer group ID - io_setup_pbuf_ring(): Sets up the physical buffer ring region and handles memory mapping for provided buffer rings This is a preparatory change for upcoming kernel-managed buffer ring support which will need to reuse some of these helpers. Signed-off-by: Joanne Koong --- io_uring/kbuf.c | 129 +++++++++++++++++++++++++++++++----------------- 1 file changed, 85 insertions(+), 44 deletions(-) diff --git a/io_uring/kbuf.c b/io_uring/kbuf.c index 67d4fe576473..850b836f32ee 100644 --- a/io_uring/kbuf.c +++ b/io_uring/kbuf.c @@ -596,55 +596,73 @@ int io_manage_buffers_legacy(struct io_kiocb *req, unsigned int issue_flags) return IOU_COMPLETE; } -int io_register_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg) +static int io_copy_and_validate_buf_reg(const void __user *arg, + struct io_uring_buf_reg *reg, + unsigned int permitted_flags) { - struct io_uring_buf_reg reg; - struct io_buffer_list *bl; - struct io_uring_region_desc rd; - struct io_uring_buf_ring *br; - unsigned long mmap_offset; - unsigned long ring_size; - int ret; - - lockdep_assert_held(&ctx->uring_lock); - - if (copy_from_user(®, arg, sizeof(reg))) + if (copy_from_user(reg, arg, sizeof(*reg))) return -EFAULT; - if (!mem_is_zero(reg.resv, sizeof(reg.resv))) + + if (!mem_is_zero(reg->resv, sizeof(reg->resv))) return -EINVAL; - if (reg.flags & ~(IOU_PBUF_RING_MMAP | IOU_PBUF_RING_INC)) + if (reg->flags & ~permitted_flags) return -EINVAL; - if (!is_power_of_2(reg.ring_entries)) + if (!is_power_of_2(reg->ring_entries)) return -EINVAL; /* cannot disambiguate full vs empty due to head/tail size */ - if (reg.ring_entries >= 65536) + if (reg->ring_entries >= 65536) return -EINVAL; + return 0; +} - bl = io_buffer_get_list(ctx, reg.bgid); - if (bl) { +static struct io_buffer_list * +io_alloc_new_buffer_list(struct io_ring_ctx *ctx, + const struct io_uring_buf_reg *reg) +{ + struct io_buffer_list *list; + + list = io_buffer_get_list(ctx, reg->bgid); + if (list) { /* if mapped buffer ring OR classic exists, don't allow */ - if (bl->flags & IOBL_BUF_RING || !list_empty(&bl->buf_list)) - return -EEXIST; - io_destroy_bl(ctx, bl); + if (list->flags & IOBL_BUF_RING || !list_empty(&list->buf_list)) + return ERR_PTR(-EEXIST); + io_destroy_bl(ctx, list); } - bl = kzalloc(sizeof(*bl), GFP_KERNEL_ACCOUNT); - if (!bl) - return -ENOMEM; + list = kzalloc(sizeof(*list), GFP_KERNEL_ACCOUNT); + if (!list) + return ERR_PTR(-ENOMEM); + + list->nr_entries = reg->ring_entries; + list->mask = reg->ring_entries - 1; + list->flags = IOBL_BUF_RING; + + return list; +} + +static int io_setup_pbuf_ring(struct io_ring_ctx *ctx, + const struct io_uring_buf_reg *reg, + struct io_buffer_list *bl) +{ + struct io_uring_region_desc rd; + unsigned long mmap_offset; + unsigned long ring_size; + int ret; - mmap_offset = (unsigned long)reg.bgid << IORING_OFF_PBUF_SHIFT; - ring_size = flex_array_size(br, bufs, reg.ring_entries); + mmap_offset = (unsigned long)reg->bgid << IORING_OFF_PBUF_SHIFT; + ring_size = flex_array_size(bl->buf_ring, bufs, reg->ring_entries); memset(&rd, 0, sizeof(rd)); rd.size = PAGE_ALIGN(ring_size); - if (!(reg.flags & IOU_PBUF_RING_MMAP)) { - rd.user_addr = reg.ring_addr; + if (!(reg->flags & IOU_PBUF_RING_MMAP)) { + rd.user_addr = reg->ring_addr; rd.flags |= IORING_MEM_REGION_TYPE_USER; } + ret = io_create_region(ctx, &bl->region, &rd, mmap_offset); if (ret) - goto fail; - br = io_region_get_ptr(&bl->region); + return ret; + bl->buf_ring = io_region_get_ptr(&bl->region); #ifdef SHM_COLOUR /* @@ -656,25 +674,48 @@ int io_register_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg) * should use IOU_PBUF_RING_MMAP instead, and liburing will handle * this transparently. */ - if (!(reg.flags & IOU_PBUF_RING_MMAP) && - ((reg.ring_addr | (unsigned long)br) & (SHM_COLOUR - 1))) { - ret = -EINVAL; - goto fail; + if (!(reg->flags & IOU_PBUF_RING_MMAP) && + ((reg->ring_addr | (unsigned long)bl->buf_ring) & + (SHM_COLOUR - 1))) { + io_free_region(ctx->user, &bl->region); + return -EINVAL; } #endif - bl->nr_entries = reg.ring_entries; - bl->mask = reg.ring_entries - 1; - bl->flags |= IOBL_BUF_RING; - bl->buf_ring = br; - if (reg.flags & IOU_PBUF_RING_INC) + if (reg->flags & IOU_PBUF_RING_INC) bl->flags |= IOBL_INC; + + return 0; +} + +int io_register_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg) +{ + unsigned int permitted_flags; + struct io_uring_buf_reg reg; + struct io_buffer_list *bl; + int ret; + + lockdep_assert_held(&ctx->uring_lock); + + permitted_flags = IOU_PBUF_RING_MMAP | IOU_PBUF_RING_INC; + ret = io_copy_and_validate_buf_reg(arg, ®, permitted_flags); + if (ret) + return ret; + + bl = io_alloc_new_buffer_list(ctx, ®); + if (IS_ERR(bl)) + return PTR_ERR(bl); + + ret = io_setup_pbuf_ring(ctx, ®, bl); + if (ret) { + kfree(bl); + return ret; + } + ret = io_buffer_add_list(ctx, bl, reg.bgid); - if (!ret) - return 0; -fail: - io_free_region(ctx->user, &bl->region); - kfree(bl); + if (ret) + io_put_bl(ctx, bl); + return ret; } -- 2.47.3