1 #include <linux/kernel.h>
2 #include <linux/errno.h>
3 #include <linux/file.h>
4 #include <linux/slab.h>
6 #include <linux/io_uring.h>
12 static void __io_notif_complete_tw(struct io_kiocb *notif, bool *locked)
14 struct io_notif_data *nd = io_notif_to_data(notif);
15 struct io_ring_ctx *ctx = notif->ctx;
17 if (nd->account_pages && ctx->user) {
18 __io_unaccount_mem(ctx->user, nd->account_pages);
19 nd->account_pages = 0;
22 if (nd->zc_report && (nd->zc_copied || !nd->zc_used))
23 notif->cqe.res |= IORING_NOTIF_USAGE_ZC_COPIED;
25 io_req_task_complete(notif, locked);
28 static void io_uring_tx_zerocopy_callback(struct sk_buff *skb,
29 struct ubuf_info *uarg,
32 struct io_notif_data *nd = container_of(uarg, struct io_notif_data, uarg);
33 struct io_kiocb *notif = cmd_to_io_kiocb(nd);
36 if (success && !nd->zc_used && skb)
37 WRITE_ONCE(nd->zc_used, true);
38 else if (!success && !nd->zc_copied)
39 WRITE_ONCE(nd->zc_copied, true);
42 if (refcount_dec_and_test(&uarg->refcnt)) {
43 notif->io_task_work.func = __io_notif_complete_tw;
44 io_req_task_work_add(notif);
48 struct io_kiocb *io_alloc_notif(struct io_ring_ctx *ctx)
49 __must_hold(&ctx->uring_lock)
51 struct io_kiocb *notif;
52 struct io_notif_data *nd;
54 if (unlikely(!io_alloc_req_refill(ctx)))
56 notif = io_alloc_req(ctx);
57 notif->opcode = IORING_OP_NOP;
60 notif->task = current;
62 notif->rsrc_node = NULL;
63 io_req_set_rsrc_node(notif, ctx, 0);
65 nd = io_notif_to_data(notif);
66 nd->account_pages = 0;
67 nd->uarg.flags = SKBFL_ZEROCOPY_FRAG | SKBFL_DONT_ORPHAN;
68 nd->uarg.callback = io_uring_tx_zerocopy_callback;
69 nd->zc_report = nd->zc_used = nd->zc_copied = false;
70 refcount_set(&nd->uarg.refcnt, 1);
74 void io_notif_flush(struct io_kiocb *notif)
75 __must_hold(&slot->notif->ctx->uring_lock)
77 struct io_notif_data *nd = io_notif_to_data(notif);
79 /* drop slot's master ref */
80 if (refcount_dec_and_test(&nd->uarg.refcnt)) {
81 notif->io_task_work.func = __io_notif_complete_tw;
82 io_req_task_work_add(notif);