Lines Matching refs:bgid

26 	__u32				bgid;  member
32 unsigned int bgid) in io_buffer_get_list() argument
34 if (ctx->io_bl && bgid < BGID_ARRAY) in io_buffer_get_list()
35 return &ctx->io_bl[bgid]; in io_buffer_get_list()
37 return xa_load(&ctx->io_bl_xa, bgid); in io_buffer_get_list()
41 struct io_buffer_list *bl, unsigned int bgid) in io_buffer_add_list() argument
43 bl->bgid = bgid; in io_buffer_add_list()
44 if (bgid < BGID_ARRAY) in io_buffer_add_list()
47 return xa_err(xa_store(&ctx->io_bl_xa, bgid, bl, GFP_KERNEL)); in io_buffer_add_list()
68 bl = io_buffer_get_list(ctx, buf->bgid); in io_kbuf_recycle_legacy()
71 req->buf_index = buf->bgid; in io_kbuf_recycle_legacy()
202 ctx->io_bl[i].bgid = i; in io_init_bl_list()
259 xa_erase(&ctx->io_bl_xa, bl->bgid); in io_destroy_buffers()
288 p->bgid = READ_ONCE(sqe->buf_group); in io_remove_buffers_prep()
302 bl = io_buffer_get_list(ctx, p->bgid); in io_remove_buffers()
342 p->bgid = READ_ONCE(sqe->buf_group); in io_provide_buffers_prep()
412 buf->bgid = pbuf->bgid; in io_add_buffers()
430 if (unlikely(p->bgid < BGID_ARRAY && !ctx->io_bl)) { in io_provide_buffers()
436 bl = io_buffer_get_list(ctx, p->bgid); in io_provide_buffers()
444 ret = io_buffer_add_list(ctx, bl, p->bgid); in io_provide_buffers()
490 if (unlikely(reg.bgid < BGID_ARRAY && !ctx->io_bl)) { in io_register_pbuf_ring()
496 bl = io_buffer_get_list(ctx, reg.bgid); in io_register_pbuf_ring()
521 io_buffer_add_list(ctx, bl, reg.bgid); in io_register_pbuf_ring()
535 bl = io_buffer_get_list(ctx, reg.bgid); in io_unregister_pbuf_ring()
542 if (bl->bgid >= BGID_ARRAY) { in io_unregister_pbuf_ring()
543 xa_erase(&ctx->io_bl_xa, bl->bgid); in io_unregister_pbuf_ring()